Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
R E S E A R C H :
U N I T E D S T A T E S
United States
H. Clinton Vaughan
415-951-1786
h.clinton.vaughan@ssmb.com The SAN Book 3.0
John C. Dean
415-951-1622
(Storage Area Networks)
john.c.dean@ssmb.com
Everything You Ever Wanted to Know About
Joseph Wana
415-951-1804
Storage (but Couldn’t Find the Answer)
NAS
2
The SAN Book III – October 22, 2001
3
The SAN Book III – October 22, 2001
Table of Contents
Executive Summary .................................................................................................................................... 3
DEMAND DRIVERS................................................................................................................................ 5
State of the Storage Market ....................................................................................................................... 6
Demand Drivers: The Real World ......................................................................................................... 14
Demand Driver Analysis .......................................................................................................................... 30
Data Sheet.................................................................................................................................................. 45
SAN AND NAS..................................................................................................................................... 47
What Is a SAN?......................................................................................................................................... 48
What Is NAS?............................................................................................................................................ 62
SAN and NAS Convergence ..................................................................................................................... 67
What Is Server Clustering?...................................................................................................................... 72
SUBSYSTEMS ..................................................................................................................................... 77
Subsystems: The Heart of Storage ......................................................................................................... 78
Disk Drives to Subsystems........................................................................................................................ 82
RAID and Beyond..................................................................................................................................... 90
NAS Subsystems...................................................................................................................................... 104
RAID Is No Good Without Core Software ........................................................................................... 112
SOFTWARE ....................................................................................................................................... 133
The Magic Word Is Software.................................................................................................................. 134
Software Applications............................................................................................................................. 147
NETWORKING................................................................................................................................... 171
Storage Networking ................................................................................................................................ 172
Storage Networking I/O: HBAs and ICs............................................................................................... 178
Storage Networking Equipment ............................................................................................................ 194
PROTOCOLS ..................................................................................................................................... 225
Protocols: The Language of SANs........................................................................................................ 226
The Basics of I/O Interfaces ................................................................................................................... 231
Fibre Channel: Networking Storage .................................................................................................... 236
Fibre Channel Versus Ethernet ............................................................................................................. 239
Potential Future IP Storage Iterations.................................................................................................. 249
The Future Band: InfiniBand ............................................................................................................... 255
Recent Company Developments ............................................................................................................ 266
COVERAGE UNIVERSE..................................................................................................................... 273
Companies in Our Universe ................................................................................................................... 274
Coverage Universe .................................................................................................................................. 275
PRIVATE COMPANIES ...................................................................................................................... 281
GLOSSARY........................................................................................................................................ 367
TABLE OF FIGURES.......................................................................................................................... 391
4
The SAN Book III – October 22, 2001
Demand Drivers
5
The SAN Book III – October 22, 2001
6
The SAN Book III – October 22, 2001
60,000
Intel
Pentium
50,000
40,000
PC Units Shipped (000s)
30,000
20,000
10,000
0
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
19
19
19
19
19
19
19
19
19
19
19
19
19
19
19
19
19
19
19
Source: IDC, Bloomberg, and Salomon Smith Barney
An argument could be made that the Internet also had quite a bit to do with PC
growth in the 1990s... and we agree. So, we tried to factor this out by looking at
corporate PC growth. Figure 2 below illustrates the growth of PC shipments
between 1995 and 2000 to large institutions, whose computers are mostly attached to
LANs. These large institutions include large businesses, medium businesses, and
the government. We believe this illustration further validates our belief that SANs
will increase the use of data and points to storage growth through the advent of
SANs.
7
The SAN Book III – October 22, 2001
12,000
17.3% CAGR
10,000
8,000
6,000
4,000
2,000
0
1Q95
3Q95
1Q96
3Q96
1Q97
3Q97
1Q98
3Q98
1Q99
3Q99
1Q00
3Q00
1Q01
Source: IDC
Networking
We believe the storage Storage networking is an emerging trend that is going to unfold in the current decade
networking market is (much like the year of the LAN evolved over many years). We believe it is a very
less than 10%
penetrated, going to
important technology alternative to throwing expensive headcount at the ongoing
85%–90%. challenge of managing more data. We believe the storage networking market is less
than 10% penetrated, going to 85%–90%. In other words, you ain’t seen nothing
yet. Dataquest projects the storage networking market will grow at a 67.0% CAGR,
to $16.9 billion in 2005.
We expect Brocade to continue leading the industry’s growth with its dominant
market position and Inrange to be the market’s up-and-comer by focusing more on a
total solution in the high end with its unique direct sales force. We also believe that
McData will continue to put up solid growth with its end-to-end product focus,
covering both the high-end and low-end markets.
We expect to see excellent growth from the HBA players QLogic, which also has a
switch business that has chalked up some impressive growth, and Emulex, which
has been aggressively penetrating new market segments and customers.
Software
Storage software is an evolving and expanding market that we expect to be an
industry outperformer. We believe the storage software market is undergoing
consolidation (which we would have characterized as “rapid” if IT spending were
8
The SAN Book III – October 22, 2001
healthier). That is to say, the three leading storage software players control a
meager 57.9% of the market with unnamed white box vendors capturing 15%–20%
(EMC with 25.5%, Veritas with 16.3%, and IBM with 16.1% market shares in 2000
according to Dataquest).
We expect EMC and Veritas to continue increasing their respective positions in this
segment; and consequently, to also continue to intensify their head-to-head
competition. Also, we wouldn’t count out Legato, which has a new management
team in place and has been making strides with its partnering approach to the
marketplace.
Subsystems
Storage subsystems are at the heart of storage and we expect that they will continue
to be the major influence in the broader storage industry. We expect EMC to
continue to dominate this segment, although other high-end suppliers, such as IBM
and Hitachi Data Systems (HDS), should continue gaining traction as well. We
expect the mid- and low-end markets to continue to be led by Compaq (which is
scheduled to be acquired by Hewlett-Packard) and NetApp (the industry’s clear
NAS leader) and believe that Dell and Sun Microsystems will play more important
roles going forward.
Commodity?
We find it hard to believe We do not see storage becoming a commodity and find evidence of this at recent
that an industry with this industry events. EMC’s August 1-2, 2001 analyst day (with a full day dedicated to
much innovation and
this much end user
technology tutorials) attracted over 300 investors. Brocade’s first storage
interest in learning about networking conference on June 11–13, 2001 attracted over 500 end users and
new developments is partners. Veritas’s VISION 2000 on October 24, 2000 attracted over 2,500 people,
quickly on its way to (mostly end users and partners: up by about 1,000 from a year earlier and 500 from
becoming a commodity.
two years earlier) to learn about its new direction into storage networking.
In short, we find it hard to believe that an industry with this much innovation and
this much interest in learning about new developments is quickly on its way to
becoming a commodity. Commodity markets are the result of mature industries in
which there is very little product differentiation. We do not see storage moving to a
commodity market for many years.
Market Projections
We believe data storage is one of the most attractive long-term infrastructure plays.
According to Dataquest projections, the overall storage market for multi-user servers
is expected to grow from $36.0 billion in 2000 to $92.0 billion in 2005 in aggregate,
representing a 20.7% CAGR (2000–05).
9
The SAN Book III – October 22, 2001
CAGR
(Revenue in $millions) 1998 1999 2000 2001 2002 2003 2004 2005 2000-2005
RAID $22,287 $25,593 $29,424 $29,841 $33,176 $38,756 $47,080 $58,399 14.7%
Internal $4,762 $5,818 $7,064 $7,269 $7,223 $6,987 $6,427 $5,704 (4.2%)
JBOD $5,083 $5,704 $4,582 $3,887 $3,162 $2,459 $1,824 $1,289 (22.4%)
External $12,093 $13,491 $16,330 $16,735 $19,270 $23,963 $31,327 $41,063 20.3%
NAS $349 $580 $1,448 $1,950 $3,521 $5,347 $7,502 $10,343 48.2%
Networking $159 $518 $1,300 $2,140 $3,610 $7,360 $13,330 $16,880 67.0%
Software $2,890 $4,088 $5,252 $6,641 $8,456 $10,709 $13,396 $16,729 26.1%
Total Projection $25,336 $30,199 $35,976 $38,622 $45,242 $56,825 $73,806 $92,008 20.7%
YoY Growth -- 19.2% 19.1% 7.4% 17.1% 25.6% 29.9% 24.7% --
100,000
90,000 Storage CAGR: 18.4% Server CAGR: 6.6%
80,000
70,000
60,000
50,000
40,000
30,000
20,000
10,000
0
1999 2000 2001E 2002E 2003E 2004E
10
The SAN Book III – October 22, 2001
*
tion)
o r p rojec
i
R (pr
CAG nic &
N e w th
24.9% Orga Data Gr
ow
i on
$ Appl
ic a t
20.3%
CAG
R*
th
ata Grow
Flat to declining Organic D rw ay
unde
storage growth Recovery
Resuming
‘normal’
industry
Impact of an growth rate
economic pullback
Time
*Dataquest’s 5 year CAGR
11
The SAN Book III – October 22, 2001
Positives:
➤ In this soft economy, storage companies have not been plagued by excessive
inventory levels similar to that of other technologies (i.e., PC and server
markets).
➤ Data is still growing from mission-critical applications, the Internet, video
streaming, and e-mail. Information growth varies widely across these
applications, but sometimes tops 100% per year in mission-critical applications
like CRM, ERP, and data warehousing.
➤ Pricing of software and storage networking remains somewhat stable despite
increased competition.
➤ We’ve uncovered some large deals and believe storage networking is the next
high growth area within the storage market. We expect IT executives to direct
their limited budgets toward technology investments, such as SANs, that
generate high ROIs within a short payback period.
Negatives:
➤ Storage deals are requiring more top management sign-off, pushing out the
length of the sales cycle.
➤ In the bull market economy, organizations overbought storage capacity in
anticipation of high growth. Today, because of the economic downturn, we are
noticing companies defer storage purchases, by first using up their excess
storage capacity.
➤ We are seeing IT managers postpone storage purchases by increasing their
storage utilization rates. By using say, 85% of available storage rather than
50%, an organization can postpone its storage expenditures.
➤ Organizations are implementing tighter controls around their storage purchases
and are enforcing cost charge-backs to the business units. According to one
source, a year ago, storage purchases seemed to be unrestricted, but today,
management is asking users to order only what is necessary.
➤ Europe does not appear to be a viable source of growth for storage companies in
the short term due to economic conditions.
We believe there are three essential reasons storage is an important future spending
area for IT managers:
1 Spending is increasingly being focused on capital investments with the highest
ROI (return on investment). For example, investments in networked storage
infrastructures, such as SANs, are generating cost savings from increased
capacity utilization, reduced hardware requirements, and lower labor costs. Our
12
The SAN Book III – October 22, 2001
real world case studies confirm this. For example, Intuit witnessed ROIs ranging
from 123% to 525% on its SAN solutions, according to a KPMG consulting case
study.
2 High-end applications like ERP and CRM continue to produce valued
information. As the critical information generated by these applications build,
organizations will use data storage management to leverage information into a
competitive advantage.
3 We believe B2B expansion and corporate webification of larger, more
established companies should continue, ensuring demand for high availability
systems. Storage solutions ensure 24x7 availability of critical data, resulting in
greater customer loyalty and retention.
13
The SAN Book III – October 22, 2001
Take More
Take More Digital Pictures
Traditional & Less
Pictures Traditional
15% 50%
Source: Dataquest
After going through a few film cards and downloading them onto my personal
computer with a 20 GB (gigabyte) hard drive, I realized that at the rate I was going, I
would fill up my disk storage space in a year (I already have a lot of other
applications on my PC). At that time, I would either have to start deleting data or
pictures, or buy a new computer. As an alternative I could have replaced the hard
drive (I can easily buy up to 80 GB today), but that is quite painful and there is no
telling what would happen if I let loose with a screwdriver on my computer to pop
the lid off and start ripping out the drive and installing a new one. What would
happen to my old data? How do I transfer data from the old disk drive (out of the
system) to the new disk drive? Wouldn’t the computer have to be reconfigured? For
14
The SAN Book III – October 22, 2001
15
The SAN Book III – October 22, 2001
Source: PC Data
Conclusions
➤ Data is important and increasingly so; for example, if my pictures are this
important to me, imagine how important a company’s most important asset
(data) is.
➤ Data is growing in many new areas and the need to store it, replicate it, and back
it up is increasing exponentially.
➤ Spending is shifting toward storage.
➤ Storage is moving to be external and networked (i.e., accessible).
➤ Information technology managers want their data online and replicated (i.e.,
more than one copy).
Now Check This Out!
Snapfish.com saw its storage grow more than 28 terabytes in the first year of
business and expects it to double again in 2002. Snapfish has several competitors,
including AOL and Kodak. If each competitor witnessed data growth comparable to
what Snapfish expects, the combined storage demand would increase by at least 500
terabytes annually. That alone could fit on EMC’s top 25 customer list.
Driving the point home: if AOL’s “You’ve Got Pictures,” an online digital photo
service, gets just 5% of its customers from existing subscribers and each customer
takes roughly 25 pictures per month and stores it using AOL, we estimate that could
drive over 200 terabytes of storage demand, annually.
16
The SAN Book III – October 22, 2001
45,000 $14,000
40,000 $12,000
35,000
0 $-
1998 1999 2000 2001E 2002E 2003E 2004E 2005E
And that is just digital photo imaging; medical imaging is also an expanding storage
market. The University of California, Berkeley estimates that there are roughly 180
million chest x-rays and mammograms taken annually in the United States. At
roughly 8 MB per x-ray, that represents potential annual storage demand of over
1,400 TB. Again this annual storage demand is equivalent in size to three of EMC’s
top customers.
While we focused our discussion above on potential digital photography storage
demand, we could have chosen many other areas, including:
➤ mission-critical applications such as ERP (enterprise resource planning) and
CRM (customer relationship management)
➤ e-mail and the growing popularity of instant messaging
➤ wireless content and handheld devices, including Research In Motion pagers
➤ digital audio (digital music downloading)
➤ digital data archiving including e-mail archiving
➤ streaming video
➤ set-top boxes, Web TV and other digital TV services
We believe the conclusions we summarized above apply to all of these, and other,
emerging technologies.
17
The SAN Book III – October 22, 2001
Other
Storage
End-Users
National Climatic
Data Center (NCDC)
There is a consistent theme to what we are hearing from a diverse group of storage
end users including those shown in Figure 9 and others remaining anonymous
(which we have labeled Company X and Company Y):
1 Data is growing. The biggest drivers of data growth are mission-critical
applications, the Internet, video streaming, and e-mail with large media
attachments. Growth rates vary from 30% to 100%-plus in some of the more
critical applications.
2 Data growth, however, is slower in 2001 than 2000. A slowing economy
means fewer transactions and declining new application deployments.
Consequently, storage capacity bought in 2000 lasted longer than expected,
reducing 2001 storage expenditures. We believe information growth in 2001
could fall in the range of 50%–70% (for capacity, not revenues). Beyond 2001,
we expect data to return to its former trend of 100% growth per year.
3 Storage capacity utilization is rising. IT managers are using up their excess
storage capacity, by squeezing as much out of their IT infrastructure as possible.
This has resulted in delayed storage purchases in 2001. Storage Area Networks
(SANs) also lead to higher utilization rates because of their centralized
18
The SAN Book III – October 22, 2001
architecture. Most end users expect to maintain these higher levels of storage
utilization beyond 2001, which has a dampening (but not crushing) effect on
storage growth going forward.
4 Price per megabyte of storage is declining. End users are telling us they are
seeing (or forecasting in the case of post 2001) price declines of approximately:
Pre-2001 35%
2001 50%
After 2001 40%
Source: Salomon Smith Barney
5 SANs (storage area networks) generate high ROIs — largely from labor
savings, floor space reduction, and reduction of file server hardware and
maintenance to support that hardware. The decision to migrate to a SAN from a
DAS (direct-attached storage), however, must be driven by a strong business
purpose.
6 SANs target market is large. Several end users we talked with (many not
mentioned by name) were small businesses with less than 8 TBs of data. What
interested us, was that almost all of them were using or were planning to
implement SANs. This could mean broader SAN adoptions than most anticipate.
7 Data replication will increase in 2002 and beyond. Driven by declining
storage and bandwidth prices, and advances in software technology, we expect
demand for replication to continue to grow well into the future. (Veritas’s
Replicator and EMC’s SRDF are the two leading products in this area).
8 Companies’ information infrastructures are closely protected assets.
Throughout our discussions, we noted that an increasing number of IT managers
either could not discuss their storage environment or did not want to be
mentioned by name for competitive reasons. We find the fact storage
infrastructures are being treated as a tightly protected corporate asset, as a very
encouraging sign for future emphasis on spending in this area. Increasingly,
things such as discussing how many terabytes a company has, how they are
managing it, how they are growing it, and what they expect to do going forward
to better manage and use their information, is becoming a secret (i.e., a
competitive advantage).
Below we provide more details of our conversations with IT executives.
19
The SAN Book III – October 22, 2001
time of year is in the month leading up to the tax filing deadline and you have a
narrow window of opportunity to perform your system backups. Unfortunately, the
actual time required to do the backup keeps increasing due to system load. Just two
days before the critical April 15 tax deadline it takes 26 hours to complete the
backup. This happened to Intuit just three years ago and during those 26 hours, the
system was unavailable to its online customers.
Because of this event and the advances in technology, Intuit implemented a robust
centralized storage area network (SAN), one that would provide it with the system
reliability, availability, and scalability the company was not getting from its direct-
attached storage (DAS) infrastructure. The old DAS environment had subsystems
linked to over 1,000 servers and was difficult and inefficient to manage. KPMG
Consulting, which conducted an extensive ROI (return on investment) analysis on
Intuit, found that Intuit’s SAN ROI ranged from 123% and 525%, as shown in
Figure 10 below.
Figure 10. Intuit’s SAN ROI
Return on Payback
Solution Investment Period
Storage Consolidation 296% 4 months
Backup/Restore 123% 9 months
System Availability 525% 2 months
Source: KPMG Consulting and Brocade Communications Systems Inc.
20
The SAN Book III – October 22, 2001
This has increased the time to market for new software applications, a key
advantage in a competitive business environment.
21
The SAN Book III – October 22, 2001
Here are some details of our conversations with this company’s senior IT
executives:
➤ Backup reliability: Imagine the emotional strain on an IT department if data
backup and replication processes fail every night. This is what the company
faced in its nightly data migration to the supply chain management reporting
group. After the problems were escalated, the company looked to EMC’s
TimeFinder to perform backups quickly, frequently, and without disruption.
Today, 22 months after implementing TimeFinder, systems run virtually
uninterrupted.
➤ Data growth is slowing in 2001: Prior to 2001, the company’s storage roughly
doubled every year, but this year it expects growth to be flat, because of the
slowing economy (fewer transactions) and cost-cutting measures (no new
application deployments and a reduction of some older data). The company
expects to be more disciplined about storage and hardware spending in 2002 and
beyond.
➤ Storage spending in 2001: The organization expects to cut storage spending by
60% in 2001, as part of its cost-cutting program. To do this the company will
increase its storage capacity utilization, clean up storage files by deleting
outdated and duplicate files, and restrict storage capacity for each business unit.
➤ On the technology wish list: reliable and proven virtualization software from a
reputable storage vendor.
22
The SAN Book III – October 22, 2001
through 2010; assuming an annualized 50% growth rate. Note: This assumes no
new application deployments.
Figure 11. Annual Storage Demand for GSFC EOS Satellite System
38,443
40,000
50% CAGR (2002-2010)
30,000
20,000 17,086
10,000 5,063
1,500
0
2001 2002 2003 2004 2005 2006 2007 2008 2009 2010
➤ Disk versus tape storage: Today, roughly 15% or almost 150 terabytes of
GSFC’s data is stored on disk with the remaining 85% housed in a tape archive.
The disk storage is the primary archive and GSFC uses caching to improve
access time to critical data. As disk prices decrease, GSFC expects to migrate
more data to disk from tape, which will improve access time and allow
researchers to perform data mining/warehousing on archived data.
➤ Disk versus tape: GSFC is weighing the cost versus performance benefit of
tape, enterprise performance disk, and IDE disk (desktop). GSFC will likely use
all three solutions to manage its data economically and efficiently. For GSFC,
tape is economical, reliable, portable, and secure — great for storing data which
is accessed less frequently. Enterprise performance disk is the best-performing
but most expensive of the options. It can be used very effectively in connection
with tape, especially when disk is used only for storing critical data requiring
fast access times. Finally, GSFC is also considering using IDE disk or desktop
disk, an option that has higher performance capability than tape and is about
one-tenth of the price of enterprise disk. Of course, using three different
solutions increases the complexity of the storage infrastructure. We will
continue to monitor GSFC to see how its storage infrastructure unfolds.
➤ SAN implementation: GSFC has recently begun a SAN pilot implementation,
and is looking to the SAN to provide high speed file sharing, high availability,
and reliability. Initially GSFC will house only a small fraction of its disk data
on the SAN, but if the first phase goes as planned, GSFC expects that a large
portion of its disk-based data would eventually be migrated to the SAN. In the
SAN environment, one of GSFC’s biggest concerns is data security.
23
The SAN Book III – October 22, 2001
6,000
4,000
4,000
2,000 2,000
2,000
1 20 250 350
0
1997 2001 2003 2005
Through 2005, LLNL ➤ Storage capacity growth expected for 2001–05: Through 2005, LLNL
expects to increase its
expects to increase its disk and tape storage to 2,000 and 10,000 terabytes,
disk and tape storage to
2,000 and 10,000 representing a 68% and 50% CAGR ( 2001–05), respectively.
terabytes, representing a
68% and 50% CAGR
➤ Storage capacity utilization: In the disk environment, storage capacity
(2001-05), respectively. utilization varies between 50% and 85%, depending on the resource.
➤ Drivers of information growth: Programs simulating various physics
scenarios are generating tremendous amounts of data, roughly 95%–98% of
LLNL’s total storage usage. Maintaining various inventories is driving 159%
and 117% CAGRs (1997–2005) in LLNL’s disk and tape storage, respectively.
➤ Most significant problems in moving/sharing data: bandwidth between
multiple large computing resources, reliability of file systems, followed by
manageability.
➤ Benefits of advanced storage architecture: With the advent of NAS
(networked attached storage) and NFS (network file systems) servers, LLNL has
24
The SAN Book III – October 22, 2001
been able to provide common home directories for its scientific computing users
on all platforms. In the future, LLNL expects SAN/NAS capabilities will allow
them to have an enterprise-wide file system that scales in terms of bandwidth,
capacity, and manageability (resolving its most significant problems, see above).
LLNL plans to deploy this type of technology in a three- to four-year timeframe.
This architecture will allow LLNL to have only one copy of the simulation data
and present that data on the computational engines, visualization resources, and
long-term tape archival resources.
➤ Data storage budgets are increasing: Currently, storage spending, including
disk and tape, is about 25% of LLNL’s total IT budget. LLNL expects the
storage budget to increase in 2002 to 35%, to support a new super computing
device.
25
The SAN Book III – October 22, 2001
26
The SAN Book III – October 22, 2001
2005 expects to have replicated all of its data (projected at one petabyte) to tape
for disaster recovery purposes.
➤ Hierarchical Storage Management (HSM) Software: NCDC uses a
hierarchical storage management software to determine what data should be
deemed critical and stored to the magnetic disk drives that provide faster data
access. As the price per megabyte of magnetic disk declines, the NCDC expects
to move a greater proportion of data to disk.
Source: sfgiants@mlb.com
27
The SAN Book III – October 22, 2001
software, the Giants believe the information collected from their electronic
turnstiles, e-ticket system, and video streaming will give them a competitive
advantage, improve their win/loss record, and bolster their bottom line. With only
seven out of the 30 Major League Baseball parks coming close to the Giants’ level
of technological prowess, we think they are right. As the business value of
information becomes apparent, the other parks are likely to follow suit. At seven
terabytes of capacity each, Major League Baseball parks combined could have
storage requirements similar in size of say, Citigroup, one of EMC’s largest
customers.
Figure 14. Potential Storage Demand for Major League Baseball
250
210 220
200
Terabytes
150
100
50
7
0
Giants MLB (potential) Citigroup
Storage Demand
Source: Salomon Smith Barney
We think the Giants are a great example of how traditional old-economy businesses
are beginning to store large amounts of critical information to gain a strategic
advantage. Who’s next? The NBA? The NFL? With this in mind, we highlight a
few of the Giants’ innovative technologies that are driving their information growth.
28
The SAN Book III – October 22, 2001
Figure 15. Pacific Bell Park: E-turnstiles, 40,000 seats, and 7 Terabytes Storage Capacity
Source: sfgiants@mlb.com
In addition to the e-turnstiles, the Giants have also implemented an e-ticket system
in their luxury suite area which issues tickets after fans swipe their driver’s license
or credit cards into the system. As a result, the Giants have access to a tremendous
amount of essential demographic information on their luxury box ticket holders. For
example, the Giants can analyze the number of Dodger games a fan attended, the
preferred day of the week, and whether s/he appears to like day or evening games.
According to the Giants’ senior IT executive, “I can definitely envision a world in
the not-so-distant future where we are 90-plus% paperless.”
While there are no definite plans to expand this system on a stadium-wide basis, all
suites and season ticket holders could be using this paperless ticket system in the
next two years. At that point, the Giants could be managing demographic data on
their fans at the level of a credit card company. Enabled by their strong storage
lineup, this information would give the Giants tremendous insight into their
customer needs and would greatly expand their sales and marketing capabilities.
29
The SAN Book III – October 22, 2001
30
The SAN Book III – October 22, 2001
4.50
Inflection Point: Storage demand drops as
3.75
Storage Demand (Terabytes)
1.50
0.75
0.00
Time
31
The SAN Book III – October 22, 2001
utilization in the DAS environment, the organization must purchase more storage if
server two needs additional capacity. The SAN environment depicted on the right-
hand side of Figure 17 has 4 TB (terabytes) of data, but only requires 4.75 TB of
storage. This represents a 40% decrease from the 8 TB of storage required in the
DAS environment, which could not exceed more than 50% capacity utilization.
Figure 17. Demand Effect of Increased Storage Capacity Utilization
1 TB: 50%
utilization SAN
Clients Clients Manager
Server 1 Server 1
Disk
Subsystem 1 2TB: 100% Disk
Server 2 Server 2
Disk
Subsystem 2 1.5 TB: 75% Disk
utilization Subsystem 2
Server 3
4 TB data =
Server 3 Disk .5 TB: 25% 85%
Subsystem 3 utilization utilization
Server 4
Server 4 Disk
Subsystem 4
1) 50% utilization = 4TB data / 8TB of available 1) 85% utilization = 4TB data / 4.75 TB of available
storage capacity) storage capacity
2) Needs of disk subsystem 2 requires additional storage 2) No wasted capacity with room to grow
purchase
3) Excess capacity in arrays 1,3, & 4 cannot cover needs
Real World: In its SAN environment, Citigroup orders additional storage when
capacity reaches 90% compared to 50%–60% in its DAS environment. By simply
expanding utilization rates, Citigroup tapped roughly 35% more storage capacity
without spending a dime.
Counter-Point: Data Replication — Driving Storage Demand
Data replication refers to the duplication or “mirroring” of an organization’s
mission-critical data for disaster recovery purposes, data migration, or business
operations (i.e., development and testing). As data replication increases, so should
the demand for storage. We believe the increasing market penetration of data
replication initiatives should, in the long run, have a profound affect on the storage
32
The SAN Book III – October 22, 2001
market and should offset some of the negative demand effect of higher utilization
rates.
For some organizations, like Citigroup, regulatory rules require the maintenance of
much of the company’s mission-critical information in a disaster recovery facility to
ensure high availability. For others, like Intuit (discussed above), replication of data
ensures faster and more accurate testing of new software enhancements, since
developers can quickly access a reliable and updated copy of data.
We believe that the projected declines in storage subsystem and bandwidth prices
and the advances in data replication software are all pointing to future demand for
data replication.
Note: We explain replication in more detail in the “Software Applications” section.
3,500
3,000 28% CAGR (2000-2005)
2,500
$ Millions
2,000
1,500
1,000
500
0
1999 2000 2001 2002 2003 2004 2005
33
The SAN Book III – October 22, 2001
0.50
0.40
$/MB/Array Unit
0.30
0.20
-43% CAGR 2000-2005
0.10
0.00
1998 1999 2000 2001E 2002E 2003E 2004E 2005E
34
The SAN Book III – October 22, 2001
35
The SAN Book III – October 22, 2001
36
The SAN Book III – October 22, 2001
Figure 20. Worldwide Forecast for RAID Storage (in Terabytes Shipped)
Terabytes
shipped
7,000,000
6,000,000 CAGR 83% (2000-2005)
5,000,000
4,000,000
3,000,000
2,000,000 CAGR 78% (1998-2000)
1,000,000
-
1998 1999 2000 2001E 2002E 2003E 2004E 2005E
37
The SAN Book III – October 22, 2001
140,000
120,000
Revenue ($M)
100,000
80,000
60,000 Bus App: 12% CAGR (2000-2005)
40,000
20,000
ERP: 16% CAGR (2000-2005)
-
1998 1999 2000 2001E 2002E 2003E 2004E 2005E
Data Warehousing
The growth of data warehousing software, which combines different databases into
one, makes enterprise applications even more powerful, further driving the demand
for storage. Data warehousing provides organizations with the tools to create order
and make sense out of the chaos that can arise out of accumulated information.
Figure 22 details the projected growth of data warehousing software: to $17.0
billion in 2004 from $6.8 billion in 2000, representing a 26.1% CAGR.
38
The SAN Book III – October 22, 2001
($M)
18,000
16,000
14,000
12,000
26.1% CAGR (1999-2004)
10,000
8,000
6,000
4,000
2,000
0
1999 2000 2001E 2002E 2003E 2004E
Source: IDC (December, 2000)
39
The SAN Book III – October 22, 2001
(Billions)
40
35
30
25
Worldwide
20
15
10
North America
5
0
1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
Consistent with the Citigroup experience, more users today rely on e-mail to send
content-rich audio and video data, increasing the average size of e-mail and the
storage capacity needed to manage that information. As depicted in the figure
below, it is evident that these new applications are going to require sizable databases
fueling the demand for more storage capacity.
Figure 24. Emerging Application Growth Projections
100
Avg. Application Size (TB)
E-Medicine
50
E-Mail
Data Warehouse
Electronic Security
Video Mail
0
2000 2005 2010
40
The SAN Book III – October 22, 2001
provides solutions to these needs, through its marquee functions such as data backup
and data recovery. We believe, therefore, that e-commerce growth could drive the
demand for storage in the future.
Ease of accessibility to the Internet in the form of paid Internet access providers like
America Online (AOL) and Internet Service Providers (ISPs) like AltaVista and
NetZero has introduced an increasingly higher number of people to the Internet year
after year. According to May 2001 IDC reports, 370 million devices were used to
access the Web in 2000; a number that is expected to reach 1.3 billion by 2005 —
reflecting a 28% CAGR. E-commerce spending, as shown in Figure 25, is projected
to exceed $5.03 trillion by 2005 up from $354 million in 2000 — a CAGR of
roughly 70%.
Figure 25. Emerging Application Growth Projections
1,400 $6,000
1,200 $5,000
Number of Internet Devices
1,000
(Millions) Left Axis $4,000
28% CAGR 2000-2005
$ Billions
800
Millions
$3,000
600
$2,000
400
E-Commerce Spending
($Billions) Right Axis $1,000
200 70% CAGR 2000-2005
0 $0
2000 2001 2002 2003 2004 2005
Source: IDC (April, 2001)
41
The SAN Book III – October 22, 2001
$3.0 $2.8
$2.5 $2.3
(In $billions)
$2.0
95% CAGR (2000-2005)
$1.5 $1.2
$1.0
$0.6
$0.5 $0.3
$0.1
$0.0
2000 2001 2002 2003 2004 2005
Source: Jupiter Enterprise Streaming Model (March, 2001)
42
The SAN Book III – October 22, 2001
% of employees w/broadband
$80 100%
80%
$60
60%
$40
40%
$20
20%
$0 0%
1999 2000 2001 2002 2003 2004 2005
43
The SAN Book III – October 22, 2001
$1,000
Internet Caching/Multicasting CAGR: 109.3% (2000-2005)
$800
Storage/Hosting CAGR: 71% (2000-2005)
(in millions)
$600
$400
$200
$0
2000E 2001E 2002E 2003E 2004E 2005E
120%
100%
80%
% of US Households
VCRs
60%
Telephones
40%
PCs
0%
1900 1910 1920 1930 1940 1950 1960 1970 1980 1990
Source: FCC, RAB, NAB, NCTA, EIA, IDC, and Salomon Smith Barney
44
The SAN Book III – October 22, 2001
Data Sheet
Size Terminology
As capacity of subsystems grows more each day, the terminology grows more
exotic. Here is a quick explanation of terms you may hear in the future.
Figure 30. Size Terminology: From Byte to Yottabyte
Many forms of traditional media are being converted to digital form so they can be
stored and copied onto computer storage systems. Figure 31 describes the storage
requirements needed for various media.
Figure 31. Sizing for Storage Applications
45
The SAN Book III – October 22, 2001
46
The SAN Book III – October 22, 2001
What is a SAN?
What is NAS?
SAN and NAS Convergence
What is Server Clustering?
47
The SAN Book III – October 22, 2001
3333333333333333333CQ
What Is a SAN?
➤ A SAN (Storage Area Network) is a storage networking architecture
that enables the more efficient use of storage capacity by decoupling
server and storage connectivity with a dedicated storage network.
What Is a SAN?
A SAN (Storage Area A Storage Area Network (SAN) is a storage networking architecture that enables the
Network) is a more more efficient use of storage assets by decoupling server and storage capacities with
robust and reliable data
architecture.
a dedicated network. Much like a Local Area Network (LAN), a SAN enables
greater scalability and reliability for computer storage resources, making them less
expensive to implement and easier to manage. Traditional storage subsystems attach
directly to the back-end of servers (called Direct Attached Storage, or DAS). In a
SAN, a dedicated storage network is inserted between servers and various storage
subsystems. The storage network is generally composed of a number of switches
which interconnect servers and storage.
Figure 35 below is an example of a traditional DAS architecture.
Figure 35. Traditional DAS (Direct Attached Storage) Architecture
Tape
Library
Key Considerations
1)High cost of management
Client 1
2)Full time data availability
Server 1
often requires traversing the
Subsystem 1 LAN
Client 2
LAN 3)Does not have fast data
transmission when a server
Server 2
has reached its capacity
Client 3 Subsystem 2 4)Does not provide any-to-
any connectivity
5)Scalable at the cost of
Client 4 establishing unused capacity
Server 3 Subsystem 3
48
The SAN Book III – October 22, 2001
Key Advantages
Tape •Reduces overall cost of
Library
storage management
•High availability
Client 1
•Any-to-any connectivity
Server 1 •Frees server capacity
•Reduces LAN congestion
Subsystem 1
Client 2 •Storage not dedicated to
LAN SAN one server
•Increases data transfer
Server 2
Client 3 rates
Subsystem 2
•Enables serverless &
LANless backup
Client 4 •Heterogeneous
Server 3
connectivity (typically
Subsystem 3
between UNIX & NT)
•File sharing
49
The SAN Book III – October 22, 2001
available to the server it’s directly attached to, as is the case with a traditional
DAS. In a traditional architecture, excess capacity often goes unused.
$70,000
$40,000
$30,000
$20,000
$10,000
$0
1998 1999 2000 2001 2002 2003 2004 2005
SAN growth is expected to exceed that of DAS (which includes JBOD and external
storage not attached to a SAN) and NAS in absolute dollars. Figure 38 below
approximates relative growth of the architectures based on RAID-based disk storage.
50
The SAN Book III – October 22, 2001
$35,000
$30,000
$25,000
DAS
$20,000
$15,000
$10,000
SAN
$5,000
NAS
$0
1998 1999 2000 2001 2002 2003 2004 2005
Available
DAS SAN Today in SAN?
1. Aggregate Tape Backup No Yes Yes
2. Capacity Utilization ~50% ~85% Yes
3. Fan-In No Yes Yes
4. Centralized Location No Yes Yes
5. GB Per Person Managed <100 >500 Yes
6. Scalable architecture No Yes Yes
7. High Availability No Yes No
Source: Salomon Smith Barney
51
The SAN Book III – October 22, 2001
megabyte of disk storage. Hence tape is often used to store data that is accessed less
frequently or for archival purposes.
In a DAS environment, each server and storage subsystem must have its own
separate tape device or drive (see Figure 40 below). The main disadvantage to this
is the management and collection process. In this layout, each tape drive could be
located far away from the other. (One IT manager jokingly suggested his employees
should wear roller skates so they could collect all of the tapes located throughout the
enterprise.) Since a single-tape drive can only store around 100 GB, tapes often
have to be replaced on a daily basis and filed into a Dewey Decimal type library.
Figure 40. Tape Backup Using DAS
Client 1
Tape
Server 1 Subsystem 1 Drive 1
Client 2
LAN
Tape
Server 2 Subsystem 2 Drive 2
Client 3
Client 4 Tape
Subsystem 3
Server 3 Drive 3
Another alternative is to back up over the LAN, which we discuss in a later section
titled “The Future of SANs.”
In a SAN environment In a SAN environment, the tape backup process is improved by being centrally
the tape backup process located and thus easier to manage. The collection of subsystems on the network can
is improved by being
centrally located and
back-up their data to an individual tape library instead of to geographically-
thus easier to manage. dispersed, single-tape drives. Tape libraries are enterprise-level automated devices
that hold multiple tapes and thus more total storage. For example, StorageTek’s
PowderHorn 9310 tape library can store up to 360 TB of data (versus 100 GB in a
tape drive). Since it can hold more data than a single-tape drive, the tapes do not
have to be collected as frequently.
52
The SAN Book III – October 22, 2001
Client 1
Server 2 Subsystem 1
Client 2
LAN SAN
Server 2 Subsystem 2 Tape
Library
Client 3
Client 4
Server 2 Subsystem 3
2. Capacity Utilization
Another benefit to SANs is higher capacity utilization. In a DAS environment, each
server typically has its own dedicated subsystem. In talking with IT managers, we
have found 50% to be a typical capacity utilization rate (percentage of total available
storage being used) in these environments.
We have found 85% to In a SAN environment, multiple servers can store data on a single larger subsystem
be a typical capacity (see Figure 42 below). In a SAN environment, we have found 85% to be a typical
utilization rate in a SAN
environment versus 50%
capacity utilization rate. IT managers often allow higher utilization rates for
in a DAS environment. subsystems in a SAN environment because it is easier to monitor a single storage
unit reaching its maximum limit versus many units in a distributed DAS
environment.
Figure 42. Better Capacity Utilization in a SAN
DAS SAN
8 small subsystems, 1 large subsystem,
50% utilization each 85% utilization
8 TB of storage used 8.5 TB of storage used
16 TB of storage capacity 10 TB of storage capacity
Subsystem 1
Server 3 Subsystem 3 Server 7 Subsystem 7
Server 3
53
The SAN Book III – October 22, 2001
DAS SAN
8 servers + 8 servers +
8 smaller, mid-range subsystems 3 larger, enterprise subsystems
Server 1
Server 3
SAN
Server 2 Subsystem 2 Server 6 Subsystem 6 Server 4
Server 5
Subsystem 2
Server 6
Server 3 Subsystem 3 Server 7 Subsystem 7
Server 7
Subsystem 3
Server 8
Server 4 Subsystem 4 Server 8 Subsystem 8
4. Centralized Location
Another benefit to a SAN is the centralized location. In a DAS environment, servers
are often located in multiple locations within an enterprise. Since DAS typically
uses interconnects with a limited distance capability (e.g., Ultra 160 SCSI has a limit
of 12 meters), the storage must be located close to the server. Hence, storage
subsystems must be located close to their respective server counterparts and cannot
be aggregated.
By locating all of the In a SAN environment, the storage can be located further away from the server since
storage in one location it is networked. For example, using Fibre Channel, storage can be located 10 km
using large, dense
subsystems versus
away. Additionally, by using channel extenders (devices that extend the length at
smaller ones, less floor which data can be transmitted), storage can be located as far as 200 km. As a result,
space is taken up per all of an enterprise’s storage can be located in one central location. By locating all
megabyte. of the storage in one location using large, dense subsystems versus smaller ones, less
floor space is taken up per megabyte. Additional cost saving can be realized by
54
The SAN Book III – October 22, 2001
spreading the fixed costs of real estate (i.e., air conditioning, lighting, security) over
more gigabytes.
5. Low Cost of Management: More GBs Per Person
A SAN architecture A SAN architecture maximizes the efficient use of each IT administrator through its
maximizes the efficient user-friendly feature/functions and its design to be physically centrally located.
use of each IT
administrator.
StorageTek recently researched and measured the benefits of SANs and found that:
➤ In a distributed storage environment, each IT administrator can manage
approximately 30–100 GB of data, generating people costs representing an
estimated 55% of the overall storage budget.
➤ In a centralized storage environment, each IT administrator can manage 500 or
more GBs of data, lowering people costs to an estimated 15% of the overall
storage budget.
Figure 44. Cost Advantage of Centralized Storage
$/MB/MO
3 Cost without storage
management software
Disk Storage
2
ost
s onC
Per
1 Cost with storage
management software
0
1978 1984 1995 2000
6. Scalability
A SAN can allocate SANs scale very efficiently, which is not to say that traditional DAS does not scale.
excess capacity where The advantage of a SAN is that it can allocate excess capacity where needed. In
needed.
Figure 45 below, in a traditional DAS architecture, excess storage capacity on
Subsystem 1 would go unused if Server 1 did not need it. In a SAN architecture,
excess storage capacity on Subsystem 1 could be allocated to Server 2 if needed.
55
The SAN Book III – October 22, 2001
DAS SAN
Subsystem 1: Utilization 10% Subsystem 1: Utilization 55%
Subsystem 2: Utilization 100% Subsystem 2: Utilization 55%
Subsystem 2 is full Neither subsystem is full
SAN
Server 2 Subsystem 2
Subsystem 2
Server 2
Client 1
Server 1 Subsystem 1
Client 2
LAN SAN
Server 2 Subsystem 2
Client 3
Client 4
Server 3 Subsystem 3
56
The SAN Book III – October 22, 2001
DAS backups require the involvement of application servers and are, therefore,
interrupted when servers fail (see Figure 48 below).
Figure 48. Traditional DAS (Direct Attached Storage) Backup Stops when a Server Fails
Tape
Library
Client 1 DAS
Server 1
(Traditional Storage Backup)
Subsystem 1
•In a tradit ional DA S
Client 2
LAN architecture, server and
network availability is vital.
Server 2 •DAS backups cannot be
Client 3 Subsystem 2
completed when a server
fails.
Client 4
Server 3 Subsystem 3
57
The SAN Book III – October 22, 2001
The SAN architecture The SAN architecture creates redundancies that result in the High Availability (HA)
creates redundancies of data through any-to-any connectivity. For example, if Server 1 fails in a SAN
that result in the High
Availability (HA) of data
architecture, Subsystem 1 can still be accessed from Server 2. Notice that SANs can
through any-to-any do this without increasing LAN traffic.
connectivity.
In a SAN environment, server failures do not interrupt data access. (see Figure 49
below)
Figure 49. SANs Create Multiple Paths to Circumvent Server Failures
Tape
Library
Client 1
Server 1
SAN
•In a SAN, a failed
Subsystem 1 server can be
Client 2
LAN circumvented.
Server 2
Client 3
Subsystem 2
Client 4
Server 3
Subsystem 3
Client 1
Server 1
SAN
•In a SAN, a storage
Subsystem 1 networking failure
Client 2
LAN can be circumvented.
Server 2
Client 3
Subsystem 2
Client 4
Server 3
Subsystem 3
58
The SAN Book III – October 22, 2001
Figure 51. SANs Create Multiple Paths to Circumvent Storage Subsystem Failures
Tape
Library
Client 1
Server 1
SAN
•In a SAN, data can
Subsystem 1
be backed up on
Client 2
LAN multiple storage
subsystems and
Server 2 mirrored data can be
Client 3
Subsystem 2 accessed from
multiple storage
subsystems.
Client 4
Server 3
Subsystem 3
SAN
Manager? Tape
Library
Client 1
Server 1
Subsystem 1
Client 2
LAN SAN
Server 2
Client 3
Subsystem 2
Client 4
Server 3
Subsystem 3
59
The SAN Book III – October 22, 2001
SAN Adoption
We believe adoption Despite recent economic pressure within the industry, we are still firm believers in
rates are going to SAN adoption. We believe adoption rates are going to dramatically increase when
dramatically increase
when the economy
the economy stabilizes and IT budgets return to normal levels. That is not to say
stabilizes and IT budgets that SAN adoptions have stopped in this environment, but they have slowed along
return to normal levels. with everything else. Additionally, the availability of serverless backup and other
SAN “killer applications” should also drive future growth.
Figure 53 below illustrates the increasing adoption rate of SANs. Going forward,
the SAN architecture will represent a larger percentage of the total market based on
RAID-based disk storage growth.
60
The SAN Book III – October 22, 2001
100%
80%
DAS
60%
SAN
40%
20%
NAS
0%
1999 2000 2001 2002 2003 2004 2005
We have heard people talking about how many SANs have actually been installed.
The key premise to understand is how SAN is being defined. Some define it as
Fibre Channel revenues, while others define it as multiple servers (more than two)
being connected to multiple storage devices (more than two). Although the former
is more common, we believe the latter is more accurate.
Technology has historically been an area in which adoption rates can linger. We
estimate “the year of the SAN” will be no different; taking off over the course of the
next several years and experiencing a protracted adoption period.
In the end we expect In the end we expect SANs to move from about 5% penetration today to 90% in the
SANs to move from future. From an account penetration point of view we believe SAN penetration is
about 5% penetration
today to 90% in the
closer to 15%. But note, not all customers who have adopted SANs have rolled it
future. out into 100% of their environment. We believe even existing adopters are only
10%–30% penetrated from a potential addressable port connectivity point of view
(again, on their way to 90%).
61
The SAN Book III – October 22, 2001
What Is NAS?
➤ NAS (Network Attached Storage) is a storage networking architecture
that allows storage subsystems to connect directly onto the existing
computing network, the Local Area Network (LAN).
What Is NAS?
NAS (Network Attached NAS (Network Attached Storage) is a storage networking architecture that allows
Storage) is a storage storage subsystems to connect directly onto the existing computing network, the
networking architecture
that allows storage
Local Area Network (LAN). Traditional storage subsystems apply Direct Attach
subsystems to connect Storage (DAS) to the back end of servers and have used costly application server
directly onto the existing capacity to transfer data from storage devices to clients (desktop computers). This
Local Area Network “application server tax” is avoided in the NAS architecture by attaching a smaller,
(LAN).
less expensive server (referred to as a thin server, file server, or Filer) to a disk
subsystem and directly attaching the combined subsystem (referred to as a NAS
appliance) onto the LAN, thus forgoing the traditional dependence of file system
access on application servers (see Figure 54 below).
Figure 54. Traditional NAS Architecture
Subsystem 1
Key Advantages
•Frees server capacity
•Storage not dedicated to
Client 1
one server
Server 1
Tape •Uses existing TCP/IP
Library
Client 2 networks
LAN •Client can directly access
Subsystem 2
Server 2 data
Client 3
•Heterogeneous
connectivity (between
Client 4 UNIX & NT)
NAS •Simplifies Data
Appliance
Management
Through the NAS architecture, storage subsystems gain access to multiple hosts
(servers and desktops), as seen in Figure 54 above. For example, the NAS appliance
can speak to Client 1 or Client 2 without major application server involvement.
Figure 55 below lists major differences in NAS and SAN.
62
The SAN Book III – October 22, 2001
SAN NAS
Protocol Fibre Channel NFS, CIFS
protocol
Network Fibre Channel Ethernet
Transfer Block File
Embedded File System No Yes
Source: Salomon Smith Barney
A NAS architecture A NAS architecture lowers storage management costs by simplifying the storage
lowers storage process and achieving heterogeneous connectivity, thereby enabling file sharing
management costs by
simplifying the storage
across multiple platforms which allows storage to be centrally managed.
process and achieving Note: Unless stated explicitly, we will refer NAS to mean disk-based NAS, not
heterogeneous
connectivity. optical NAS, nor tape-based NAS.
63
The SAN Book III – October 22, 2001
64
The SAN Book III – October 22, 2001
Auspex
High-End Broadband Storage
Compaq
EMC
Hewlett Packard
IBM
Midrange Network Appliance
Network Storage Solutions
Price Procom
3Ware Sun Micro
Hewlett Packard
IBM
Entry Maxtor
Network Storage Solutions
Intel Procom
Maxtor Quantum (Connex)
Nortel
Quantum
Performance/Availability
Source: Salomon Smith Barney
Entry NAS
These are products designed to provide inexpensive, simple file sharing for smaller
workgroups, or as an alternative to portable storage technologies such as Iomega’s
external Jaz. The uses of these products may be permanent for small businesses —
for example, replacing the need for a traditional file server. Examples of products in
this class are Quantum’s Snap Servers and Maxtor’s MaxAttach.
Midrange NAS
These products are designed to have increased performance, scalability, and
reliability from Entry NAS. Pricing typically ranges from $5,000 to $100,000 with
maximum capacities scaling to 2 TB. Examples of products in this class are
Network Appliance’s F85 and Compaq’s TaskSmart N2400.
High-end NAS
Examples of NAS These products are designed for use in data centers that are expected to provide
products in the various multiple terabytes of storage to the network or application servers. The products
segments can be found
in the “NAS
must deliver high availability and scalability, and thus they are characterized as
Subsystems” section. having more than one storage controller; nearly always having redundant
components, multiple processors, and network interface cards; and usually being the
first to adopt new performance-oriented storage technologies. Pricing starts at
$100,000 with capacities scaling to 12 TB in clustered configurations. Some
examples of products in this class are EMC’s Celerra and Network Appliance’s
F800 Series.
Examples of NAS products in the various segments can be found in the “NAS
Subsystems” section.
65
The SAN Book III – October 22, 2001
While the high-end only While high-end NAS ($100,000 and above) only commanded 6% of total NAS
commanded 6% of total units, it dominates total revenues, representing 71% of total NAS revenues in 2000.
NAS units, it dominates
total revenues,
Conversely, the entry level ($25,000 and below) has approximately 85% of total
representing 71%. units but represented only 9% of total revenues in 2000 (Figure 58 below).
Figure 58. NAS Units and Projections
CAGR
Price Segmentation 1998 1999 2000 2001 2002 2003 2004 2005 2000-2005
Greater than $100,000 1,514 2,285 3,834 4,870 8,959 13,833 20,481 29,312 50.2%
$25,000 to $100,000 1,527 2,337 5,486 6,036 15,364 27,620 46,041 74,543 68.5%
$5,000 to $25,000 -- 1,083 3,852 10,727 21,325 35,636 58,772 91,650 88.5%
$2,000 to $5,000 -- 1,275 18,495 26,575 59,680 104,211 158,566 227,004 65.1%
Less than $2,000 -- 19,690 30,500 33,725 47,706 62,756 78,611 96,521 25.9%
Total RAID Revenue 3,041 26,670 62,167 81,933 153,034 244,056 362,471 519,030 52.9%
Year-to-Year Growth -- 777.1% 133.1% 31.8% 86.8% 59.5% 48.5% 43.2% --
Network Appliance, EMC, and Quantum, combined, have over 90% of the total
NAS market share. Figure 59 below illustrates NAS market share by revenue in
2000.
Figure 59. Total NAS Market Share by Revenue, 2000
Dell
2.0% Others
7.9%
Quantum
4.0%
Network Appliance
49.8%
EMC
36.3%
66
The SAN Book III – October 22, 2001
➤ In our opinion, the merging of SAN and NAS will further drive more
demand for storage networking software and equipment, similar to
the way LANs increased demand for PCs and servers.
Data-Centric Architectures
We believe SAN, NAS, We believe all debates around NAS, SAN, and traditional DAS (Direct Attached
and DAS architecture Storage) should dissipate in favor of one “Data-Centric” architecture. In a Data-Centric
have been merging and
will continue to merge.
architecture, nobody cares what you call it, whether it be SAN, NAS, Fibre Channel,
iSCSI, or Ethernet. The only thing that matters is whether or not it works! Keep in
mind, the whole reason SAN and NAS have emerged is to solve today’s data puzzle,
which is: 1) data is doubling every year, and 2) IT staffing budgets are remaining flat.
In other words, customers need to manage more data, but with no additional resources.
The other important factor to keep in mind is that nobody is going to throw out current
equipment just because newer equipment is better. Simply stated, newer equipment that
is backward-compatible with the current equipment is what customers are most likely
going to buy. This benefits everyone: 1) customers migrate toward better, more
advanced solutions, 2) customers get to reap the benefits of past investments, and 3)
new technologies have the opportunity to thrive through adoption.
67
The SAN Book III – October 22, 2001
Subsystem 1
Client 1
Server 1
Tape
Library
Client 2
LAN
Subsystem 2
Server 2
Client 3
Client 4
NAS
Appliance
To add greater tape backup capacity, another tape library must either be added to the
LAN, to the back-end of an application server, or to the back-end of the NAS (the
latter of which we would call a SAN if it interconnected multiple NAS subsystems).
To add more disk capacity, another NAS appliance must be added to the LAN or to
the back of the NAS.
Adding tape connectivity Figure 61 below illustrates how adding tape connectivity to the back-end of NAS
to the back-end of NAS develops a NAS/SAN hybrid.
develops a NAS/SAN
hybrid. Figure 61. By Scaling NAS, It Becomes a SAN
Tape
Library
Key Considerations
1)Relatively low cost of
Client 1
Server 1
management
Disk
2)Fault tolerant
Array 1 3)Data transmission speeds
Client 2 NAS Disk
LAN Appliance 1 are somewhat constrained
by the LAN
4)Heterogeneous
Client 3 N
SA connectivity
NAS Disk
5)Scalable at the cost of
NAS Tape
Appliance 2 Appliance losing its simplicity
Client 4
68
The SAN Book III – October 22, 2001
When scaling a NAS When scaling a NAS architecture, it makes the most sense to us (with today’s
architecture, it makes the technologies) to merge NAS and SAN architectures to achieve the most efficient
most sense to us (with
today’s technologies) to
results. For this reason, we believe SAN and NAS architectures complement one
merge NAS and SAN another.
architectures to achieve
the most efficient Network Appliance FIlers Are Interoperable with Brocade SANs
results. In May, 2001, Network Appliance announced that its filers were interoperable with
Brocade SANs, allowing configurations like the one depicted in Figure 61 above.
By supporting SAN-enabled backups, we believe Network Appliance opens up its
solution to greater market potential. By putting the tape backup solution on a SAN
behind NAS instead of onto the LAN, Network Appliance filers offer reduced
network traffic on the LAN, high availability of the tape appliance, high availability
of its Filers, and greater scalability, thereby creating a more robust computing
environment.
Tape
Library 1
Client 1
Server 1
Tape
Client 2 NAS Disk Disk Library 2
LAN Appliance 1 Array 1
Disk
Client 3 N
SA
Replication 1
Disk
NAS Disk Replication 2
Appliance 2
Client 4
69
The SAN Book III – October 22, 2001
Tape
Library
Key Considerations
1)Relatively low cost of
Client 1
Server 1
management
Symmetrix 1
2)Fault tolerant
Client 2
3)Data transmission speeds
LAN Celerra 1
are somewhat constrained
by the LAN
Symmetrix 2
4)Heterogeneous
Client 3
connectivity
5)Scalable at the cost of
losing its simplicity
Client 4 Celerra 2 Symmetrix 3
70
The SAN Book III – October 22, 2001
Key Considerations
1)Relatively low cost of
Client 1
Server 1
management
EMC
2)Fault tolerant
Symmetrix 1 3)Data transmission speeds
Client 2
LAN EMC
Celerra 1 are somewhat constrained
by the LAN
EMC
Symmetrix 2 4)Heterogeneous
Client 3 N
SA connectivity
EMC
5)Scalable at the cost of
Celerra 2 EMC losing its simplicity
Client 4 Traditional Symmetrix 3
NAS
Appliance
Tape
Library 2
Application
Subsystem
Client 1 Server 1
1
Subsystem 2
Client 2
LAN SAN
Application
Server 2
Client 3
Subsystem 3
71
The SAN Book III – October 22, 2001
Servers 1 & 2
Server 6 Server 3
Servers 4 & 5
Switches could also be used to enable higher node count clusters as depicted in
Figure 67 below.
72
The SAN Book III – October 22, 2001
Tape
Library 1
Client 1
Client 2 Subsystem 1
Subsystem 2
Client 3
LAN
SAN
Subsystem 3
Client 4
Subsystem 4
Client 5
Subsystem 5
Client 6 Subsystem 6
Client 7 Tape
Library 2
Subsystem 7
73
The SAN Book III – October 22, 2001
Figure 69 below illustrates another way to unite server clusters and storage
networks, potentially using multiple protocols, including FC-SCSI, FC-IP, and FC-
VI, enabling server-to-server, server-to-storage, and storage-to-storage connectivity.
This is the ultimate system area network.
Figure 69. Designing a Potential Systems Area Network (san)
s
ver
Se r
Clients
Storage
LAN
san
S er
ver
s
74
The SAN Book III – October 22, 2001
$70
$60
140% CAGR
$50
$40
$30
$20
$10
$0
1Q99 2Q99 3Q99 4Q99 1Q00 2Q00 3Q00 4Q00 1Q01 2Q01
Clustering uses the Virtual Interface (VI) and, since InfiniBand is not available
today, Fibre Channel has become the preferred networking medium to run VI due to
Fibre Channel’s efficient design. That said, IP is the most proliferated protocol for
server clustering. This has created a natural partnership between VI and Fibre
Channel which, in turn, has opened up many storage networking companies to the
development of Infiniband in order to expand their addressable market. Intel has
invested in Ancor (owned by Qlogic), Crossroads, and other emerging networking
companies to promote the development of Infiniband. Startups such as Mellanox
and Banderacom are focused on developing next-generation Infiniband technology
and have also drawn the attention of Intel.
There is currently a ➤ Ethernet is not typically used to run VI as a result of IP’s inefficient protocol
strong partnership stack. We believe this is one of the largest drivers behind many Ethernet
between VI and Fibre
Channel, which has
networking companies’ efforts to upgrade the IP stack: to be able to cluster more
resulted in the use of efficiently.
Fibre Channel to network
large server clusters. ➤ Novell, Veritas (Cluster Server, Global Cluster Manager), and Legato
(Automated Availability Manager) have each released 32-node cluster software
packages with the ability to expand. While many sans may start small (under
ten server nodes), customers are taking advantage of existing scalable clusters
which can scale from two to an almost infinite number of systems. In the
second quarter of 2001, Veritas’s Cluster Server represented over 20% of its
licensed revenues.
75
The SAN Book III – October 22, 2001
76
The SAN Book III – October 22, 2001
Subsystems
77
The SAN Book III – October 22, 2001
z
Subsystems: The Heart of Storage
➤ While all the buzz is about newer growth areas of storage networking
and software, the reality is that subsystems are the heart of storage.
Storage
Software
22.2%
SAN
Networking RAID Storage
13.3% 64.5%
As subsystems and As subsystems and architectures improve in functionality, performance, and price,
architectures improve in we believe more end users will invest in SANs. As SAN penetration improves, the
functionality,
performance, and price,
adoption of complementary technologies such as storage networking and software
we believe more end should offer exceptional growth.
users will invest in
SANs. Figure 72 below illustrates the high growth that storage subsystems are expected to
experience.
78
The SAN Book III – October 22, 2001
CAGR
1998 1999 2000 2001 2002 2003 2004 2005 2000-2005
Internal $4,762 $5,818 $7,064 $7,269 $7,223 $6,987 $6,427 $5,704 (4.2%)
JBOD $5,083 $5,704 $4,582 $3,887 $3,162 $2,459 $1,824 $1,289 (22.4%)
External (Subsystems) $12,093 $13,491 $16,330 $16,735 $19,270 $23,963 $31,327 $41,063 20.3%
NAS $349 $580 $1,448 $1,950 $3,521 $5,347 $7,502 $10,343 48.2%
Total RAID Revenue $22,287 $25,593 $29,424 $29,841 $33,176 $38,756 $47,080 $58,399 14.7%
Year-to-Year Growth -- 14.8% 15.0% 1.4% 11.2% 16.8% 21.5% 24.0%
1998 2000
Compaq
20.0%
Note: We have aggregated Hitachi and Hitachi Data Systems (HDS) with Hewlett-
Packard since the company OEMs HDS’s Lightning subsystem.
79
The SAN Book III – October 22, 2001
Since storage capacity is We believe the ability to understand the fine nuances of new subsystem technology
at the heart of the developments can provide an edge in helping predict a company’s future success.
industry, the subsystems
companies have
Also, since storage capacity is at the heart of the industry, the subsystems companies
tremendous influence on have tremendous influence on the direction of related technologies.
the direction of related
technologies. For example, in June, 1995, EMC introduced Symmetrix 3000, one of the world’s
first platform-independent storage subsystems, which included EMC’s first entrance
onto open systems platforms (i.e., NT and UNIX) in addition to its mainframe
platform. Although the significance of this technological development may seem
obvious in hindsight, it took about a year before the new product translated into
significant revenues since it had to target a new open system customer base. This
new product cycle fueled growth for EMC for over five years and forever changed
the dynamics in the storage industry (see Figure 74).
Figure 74. EMC’s Stock Price Versus the S&P
6000%
EMC
S&P
5000%
Platform Sales
independent product ramp up
release
4000%
Time to ramp selling to
Relative Performance
3000%
2000%
1000%
0%
Ju 4
Ju 5
Ju 6
Ju 7
Ju 8
Ju 9
Se 3
D 3
Se 4
D 94
Se 5
D 5
Se 6
D 6
Se 7
D 7
Se 8
D 8
Se 9
D 9
M 3
M 4
M 5
M 6
M 7
M 8
9
-9
-9
-9
-9
-9
-9
-9
9
-9
-9
-9
-9
-9
-9
-9
p-
n-
p-
n-
p-
n-
p-
n-
p-
n-
p-
n-
p-
ar
ar
ar
ar
ar
ar
n
ec
ec
ec
ec
ec
ec
ec
Ju
80
The SAN Book III – October 22, 2001
81
The SAN Book III – October 22, 2001
82
The SAN Book III – October 22, 2001
While disk drives vary, they all contain the same basic components. In many ways,
disk drive components are analogous to those in an audio record player: the disk
platter is analogous to the record, the head is analogous to the needle, and the
suspension/head arm is analogous to the record player arm.
Like an audio record, information is recorded on both sides of the platter. However,
in a disk drive the disk platter is fixed in place. Thus, there is a head (i.e., a needle)
on each side of the disk platter (i.e., record). Also, in a disk drive there is typically
more than one disk platter per drive. Since each disk platter can contain a fixed
amount of data, more disk platters can be stacked into a disk drive to expand its
storage capacity. Disk drives can have up to 12 platters and can be re-recorded over
(much like a tape cassette).
The disk drive in Figure 75 above has three media and six heads. Assuming it stores
33.3 GB on each platter, it will hold 100 GB of information.
Disk drives typically In Figure 76 below, to the left is what a single disk drive looks like as an internal
account for about 40% to component and to the right what it looks like as a stand-alone external product. Disk
60% of a storage
subsystem’s cost.
drives typically account for about 40% to 60% of a storage subsystem’s cost. Note:
While the disk drive on the left displays its internal architecture, disk drives in use
are fully enclosed to look like a small rectangular pizza box.
Figure 76. Disk Drives
Internal External
Source: Maxtor
Figure 77 below has an example of where a disk drive could be located within a
subsystem. In this case, we use EMC’s Symmetrix as an example of a subsystem.
83
The SAN Book III – October 22, 2001
Fans
(outtake)
RAID
(Intelligence)
Disk
Drives
Control
Console
Power
Logic and
Memory Boards
Fans
(intake)
84
The SAN Book III – October 22, 2001
JBOD Configuration
Network Subsystem
Subsystem
Server vendors have Server vendors are typically more motivated to promote internal versus external
several advantages when storage since they have several advantages when it comes to selling storage inside
it comes to selling
storage inside the server
the server including brand name familiarity and initial point of contact with the
including brand name customer.
familiarity and initial
point of contact with the Figure 79 below depicts internal versus external storage. The thick line connecting
customer. server and storage in the external configuration can represent a SCSI or Fibre
Channel cable. More complex configurations can be created with switches for
networked storage.
Figure 79. Internal Versus External Storage
Internal External
Storage
The obvious result is that many of the top internal storage vendors are also server
vendors, as illustrated in Figure 80 below.
85
The SAN Book III – October 22, 2001
1998 2000
Fujitsu, Fujitsu
Siemens
Fujitsu, Fujitsu 2.1%
Siemens Other Other
Compaq NEC
4.6% 15.4% 17.1% Compaq
29.3% 1.4%
30.3%
NEC
Hewlett Packard
2.0%
7.1%
Hewlett Packard
7.5% Sun
Sun 7.6%
Dell IBM Dell IBM
8.1%
7.6% 25.5% 12.3% 22.2%
When players such as The server vendors were able to retain a large percentage of storage market share for
EMC and StorageTek years due to its internal design. However, when players such as EMC and
entered the storage
market and
StorageTek entered the market and demonstrated the benefits of external storage, a
demonstrated the new segment emerged. Some of the benefits of external storage include better
benefits of external scalability, functionality, reliability, flexibility, serviceability, and less CPU cycles.
storage, a new segment Unlike the server vendors, storage vendors, such as EMC, were more motivated to
emerged.
sell external storage since this was the only type of storage they sold, rather than
keeping it captive to their server sale by bringing it inside the server (or “internal”).
Many server vendors evolved by offering JBOD. But while this appeared to be
“external” storage, it was not. Server vendors succeeded to some degree with this
initiative until the advent of storage area networking (SAN), which highlighted the
benefits of external storage. As a result of its numerous advantages, external storage
has been growing quickly while internal storage and JBOD revenues have begun to
decline.
From 1997 to 2000, EMC’s organic revenues grew at a compound annual growth
rate (CAGR) of 37.9%. Since it is the clear leader and only participates in the
external storage subsystems market, it is a great real-world illustration of the growth
of its market.
Figure 81 below demonstrates EMC’s Symmetrix revenue stream — with a CAGR
of 26.3% from 1995 to 2000 — as a great example of the success of external
subsystems.
86
The SAN Book III – October 22, 2001
Revenue in millions
$1,600
$1,400
CAGR 26.3%
$1,200
$1,000
$800
$600
$400
$200
$0
1Q95
3Q95
1Q96
3Q96
1Q97
3Q97
1Q98
3Q98
1Q99
3Q99
1Q00
3Q00
1Q01
Source: Company reports and Salomon Smith Barney. CAGR is based on annual revenues from 1995 to 2000 and does not include 2001.
Figure 82 below demonstrates that although the major server vendors own the
internal storage market, EMC and Hitachi Data Systems (pure play storage vendors)
are the market leaders in the external storage market. Note: We have aggregated
Hewlett Packard with Hitachi Data Systems and Hitachi since HP OEMs Hitachi’s
product.
Figure 82. Internal Storage Versus External Storage Market Share by Revenues (in 2000)
Sun
7.6% Sun Network Storage Compaq Storage
Dell IBM 7.1% IBM 11.3%
Hewlett Packard,
12.3% 22.2% 8.4%
Hitachi and HDS
17.2%
However, server vendors still have an advantage to selling and a desire to sell
JBOD. Figure 83 below illustrates that for JBOD storage, the storage is captive (the
brand of the JBOD is the same as the brand of the server) 99% of the time.
However, external storage subsystems are captive in less than half of the cases and
declining.
87
The SAN Book III – October 22, 2001
Noncaptive
0.6%
Captive
44.4%
Noncaptive
55.6%
Captive
99.4%
$45,000
$40,000
$35,000
$30,000
External
$25,000
CAGR 20.3%
$20,000
$15,000
Internal
$10,000
CAGR (4.2%)
$5,000
JBOD, CAGR (22.4%)
$0
1998 1999 2000 2001 2002 2003 2004 2005
Since server vendors have less of an advantage selling external storage than they do
internal storage, market share position can change quickly. Figure 85 shows the
market share for external storage for 1998 and 2000. Market share percentages
change more in external storage than in internal storage (see Figure 85 below). The
more dynamic nature of the external market gives independent vendors such as LSI
Logic and startups a better opportunity to penetrate the market.
88
The SAN Book III – October 22, 2001
Figure 85. External RAID Storage Market Share by Revenues (1998 and 2000)
Fujitsu
Compaq Storage
7%
8%
Hitachi, HDS Sun Network Storage Compaq Storage
IBM IBM
13% 7.1% Hewlett Packard, 11.3%
11% 8.4%
Hitachi and HDS
Sun Network Storage
Hewlett Packard 17.2%
6%
4%
89
The SAN Book III – October 22, 2001
RAID-ing Scale of 1 to 5
Which RAID algorithm a In 1987, researchers David Patterson, Garth Gibson, and Randy Katz published a
subsystem vendor paper that outlined the concept of RAID (Redundant Array of Independent Disks).
decides to offer can
impact the design, cost,
RAID is an algorithm (formula) that describes a method of writing data to a group of
reliability, and speed of disks to improve performance and protect against disk failure. Which RAID
its subsystem. algorithm a subsystem vendor decides to offer can impact the design, cost,
reliability, and speed of its subsystem. For example, a new subsystem designed
using RAID 5 would likely have greater capacity but slightly less speed than a
subsystem using RAID 1. EMC’s RAID expertise has helped propel it to the leading
position within the market. In this section we describe what these terms mean, the
different RAID levels available, and their relative advantages and disadvantages.
What Is RAID?
A subsystem with RAID intelligence inside is smarter, more reliable, and more
expensive than JBOD (Just a Bunch Of Disks) which separates the intelligence (i.e.,
RAID) from the disk. A RAID subsystem uses algorithms to write data to a group
of disks so that the drives’ combined performance and reliability are better than if
they were working independently. Several RAID algorithms exist, each with its
own advantages. The methods differ primarily in the way they write/retrieve data
and build in fault tolerance. We will explain two of the more popular methods in
detail, RAID 1 and RAID 5, and then touch upon a few others.
RAID 1: Mirroring
Mirroring is more fault- In the RAID 1 algorithm, data is written to disk using mirroring. In mirroring, every
tolerant, but takes up a time data is written to disk, an exact copy is made to a different disk in the same
lot of disk space.
subsystem (Figure 86 below). This is accomplished by writing the data to two
different drives at the same time.
90
The SAN Book III – October 22, 2001
123 123
If one drive fails, an exact copy exists on a separate healthy drive and no data is lost.
When the original failed drive is replaced with a new drive, the mirrored drive (the
backup copy) becomes the primary copy and the new drive becomes the backup
(Figure 87 below). Therefore, the subsystem always has two copies of the data,
thereby ensuing that data is not lost.
Figure 87. Redundancy in RAID 1
X 123
123
Co
py
123
Replacement Disk
In addition to improving fault tolerance, RAID 1 also enables faster read times. If
two computers are trying to access the same file simultaneously, rather than waiting
in queue, they can each access a different copy of the data. But, there are no free
lunches. The main disadvantage to RAID 1 is that it is expensive since it requires
twice the storage capacity of a non-mirrored configuration.
91
The SAN Book III – October 22, 2001
No Striping Striping
123 123
123 1 2 3
RAID 5 builds in fault RAID 5 builds in fault tolerance using fewer disks than RAID 1. The equivalent of
tolerance using fewer one disk of storage is used to store parity per each RAID set of disks.
disks than RAID 1. The
equivalent of one disk of Figure 89 below demonstrates how parity works. The RAID controller or software
storage is used to store calculates the data in the first row (the numbers 1, 2, and 3) add up to 6, which is
parity per each RAID set
of disks. then stored on a different disk. Note: In actual practice RAID 5 uses 1s and 0s for
the data, the parity calculation is stored across several disks instead of one, and the
parity calculates if the data on the other disks add up to an odd or even number.
Figure 89. Parity in RAID 5
1 + 2 + 3 = 6
Data Parity
In RAID 5, if one of the disks fails, information from the other disks can be used to
recreate the data on a spare disk. In Figure 90 below, according to the parity disk,
the data on the other disks should add up to six. By subtracting one and three from
92
The SAN Book III – October 22, 2001
six, the RAID 5 intelligence can determine a two was on the failed disk and thus
recreate it.
Figure 90. Redundancy in RAID 5
X + 0 + 3 = 6
(=6-1-3)
Replacement Disk
The benefits of RAID 5 are that it can decrease the time it takes to write since it
spreads the workload across drives and it is less expensive since it takes up less disk
space than RAID 1, but again, there are no free lunches. It also requires greater
CPU utilization in the subsystem, thereby requiring more silicon due to the greater
level of difficulty in calculating the more complex RAID 5 algorithm versus a RAID
1 algorithm. It also takes longer to rebuild due to the calculations involved.
RAID 5, however, Is RAID 5 as reliable as RAID 1? No. In a situation in which multiple drives fail at
requires greater CPU the same time, RAID 5 can actually lose data. For example, in the rare event that
utilization in the
subsystem and takes
drives 1 and 2 went down simultaneously before either could be replaced, parity
longer to rebuild. cannot be used to calculate what data was stored on each drive.
Other RAID Levels
RAID levels 2, 3, and 4 also stripe and use parity schemes slightly different than in
RAID 5. RAID 0+1 and RAID 10 use both striping and mirroring. These RAID
levels are not used as often as RAID 1 or 5. Also, EMC uses its proprietary RAID S
for its Symmetrix flagship subsystem. In RAID S, part of the RAID calculation is
performed by the disk drives themselves instead of by the disk controller. RAID S
does not use striping, but it is often compared with RAID 5. Many vendors such as
EMC provide customers with choices of multiple RAID levels.
Figure 91 below charts what RAID levels certain vendors incorporate in their
subsystems.
93
The SAN Book III – October 22, 2001
1 2 3 4 5 S 0+1 10
Compaq StorageWorks 12K FC • • • •
Dell PowerVault 660F • • • •
EMC Clariion • • • •
EMC Symmetrix • • •
HDS Lightning 9900 • •
HDS Thunder 9200 • • •
IBM ESS (Shark) •
IBM MSS • • • •
LSI Logic e4400 • • • •
Network Appliance •
Sun StorEdge T3 • •
Source: Salomon Smith Barney
94
The SAN Book III – October 22, 2001
1 2&3 4 5 6
95
The SAN Book III – October 22, 2001
5 Disk Controller: Next, the data goes to the disk controller. The disk controller
is a set of chips that contains the intelligence to manage the actual writing and
reading to disk.
6 Disk Drive: Next, the data is stored in the disk drive. Once the data is found
and retrieved, it will be sent to the host that initially requested it in nearly the
reverse order.
Subsystems use Subsystems use different combinations and architectures of hardware and software
different combinations to execute similar functions. Figure 93 below provides a different view. It gives an
and architectures of
hardware and software
idea of how various subsystems, which can be largely composed of off-the-shelf
to execute similar components, can still be unique in terms of performance and design.
functions.
Figure 93. Subsystems Architecture Model
microprocessor,
(2) Logic
algorithm, firmware
(6)
Disk SCSI, FC, ATA
96
The SAN Book III – October 22, 2001
6 Disks: A vendor can choose among different interfaces (i.e., Fibre Channel,
SCSI), speeds, and capacities.
In any design, there are There are many considerations to choose from when designing a subsystem. In any
trade-offs to using one design, there are trade-offs to using one method versus another to fulfill a function.
method versus another
to fulfill a function.
One of the more common trade-offs is price versus performance. For example, a
vendor could use the best hardware available to design a subsystem, but it would
likely be expensive to manufacture. Additionally it still might not perform as well
as an older system which had more advanced firmware.
Another example of a subsystem consideration is “to make” versus “buy.” For
example, Hewlett-Packard OEMs its high-end subsystem from Hitachi Data
Systems, allowing it a faster time to market. On the other hand, IBM has chosen to
design its own subsystem, the “Shark” (although in the past it has OEMed from
StorageTek), giving it higher potential margins and greater control over future
product releases.
Although knowing the specifications of a new subsystem can help predict a
subsystem’s performance relative to competitors, the ultimate test is still customer
adoption and sales. Customers choose vendors on many factors in addition to
“speeds and feeds” such as support, software applications, price, reliability, ease-of-
use, functionality, scalability, size, interoperability, and total cost of ownership
(TCO). Since different customers have different needs, companies that can satisfy
the broadest set of needs in the most complete way should gain the most market
share.
What Is Cache?
Cache (pronounced cash) stores data in silicon instead of magnetic drives. As a
result, it is faster and more expensive. A megabyte of storage capacity on a hard
disk drive can cost around $0.03, but on silicon such as SRAM, can cost $10. Think
of it as short-term memory (cache) versus long-term memory (disk drive). Data in
cache is easier and faster to retrieve but is not stored as long as data in disk drives.
Subsystems use cache to store the most frequently accessed data so the total amount
of time it takes to access data decreases. The amount of time it takes to access cache
is measured is nanoseconds (billionths of a second) whereas the amount of time it
takes to access disk drives is measured in milliseconds (thousandths of a second).
In addition to having a faster access time versus disk drives, getting data from cache
will also shorten the path the request must make in the subsystem before retrieving
the data. If the subsystem looks for the data and the cache already has it, the
97
The SAN Book III – October 22, 2001
subsystem can retrieve the data directly from cache without ever having to access
the disk drive. However, if it does not find the data in the cache, the subsystem must
then go to the disk drive.
Figure 94 below illustrates a cache hit (requested data is in the cache) versus a cache
miss (requested data is not in the cache).
Figure 94. Cache Hit Versus Cache Miss
1
CPU Cache Disk
2
Cache Miss: Need to access disk
1 2
CPU Cache Disk
4 3
Source: Salomon Smith Barney
Cache is analogous to closet space. In order to save time and energy, you would
probably store the outfits you wear most often in the closet nearest to you, the
bedroom closet (cache), rather than one further away, the attic closet (disk). Now
imagine that on a subsystem scale — you would only be able to store three shirts out
of 100 in that nearby closet (analogous to 32 gigabytes out of one terabyte).
There are several ways to improve the use of cache including 1) increasing the size
of the cache, 2) improving the algorithm used to decide which data is stored in
cache, 3) mirroring, and 4) battery backup. Increasing the size of the cache is
analogous to having a bigger bedroom closet. The more you can store in the
bedroom closet (cache), the less often you have to go to the attic (disk). Most
current subsystems have cache sizes ranging from several megabytes to, for
example, 64 GB on EMC’s Symmetrix. We will describe the other three variables
in greater detail next.
Speed Through Caching Algorithms
A caching algorithm is a This is one of EMC’s greatest competitive advantages, in our view. A caching
formula that tells the algorithm is a formula that tells the subsystem which data to store in cache. If the
subsystem which data to
store in cache.
cache is one gigabyte in size, the algorithm’s goal is to correctly guess which
gigabyte of data will be requested next. Returning to the closet analogy, the caching
algorithm is similar to the way you guess which clothes you will likely wear next
week so you can determine what to put in the bedroom closet. Well-constructed
algorithms can speed up the overall subsystem by maximizing the amount of data
that can be retrieved from the cache (bedroom closet) so there is less time needed to
98
The SAN Book III – October 22, 2001
access the slower disk drives (attic closet). Note: The selection of which data is
stored in the cache changes continuously.
There are two types of There are two types of algorithms: read and write. Read algorithms determine what
algorithms: read and to store in the cache. Write algorithms determine when to write the data from cache
write.
to disk. A simple example of a read algorithm is “store the data that you used most
recently in the cache.” Good caching algorithms are incredibly difficult to create
since they are essentially trying to predict the future.
Figures 95 below shows how the cache relates to accessing files.
Figure 95. Algorithms Guess Which Files Will Be Accessed Soon
Effective Ineffective
99
The SAN Book III – October 22, 2001
Data Data
X
Cache
X
Cache
Mirror
Cache
Subsystem Subsystem
A second risk to data loss on the cache is power. The type of memory used for
cache (often SRAM) is usually volatile. In volatile memory, when there is a loss of
power, the data is lost forever. This can happen if there is a power blackout, the
subsystem power supply system fails or maybe somebody walks by a subsystem and
accidentally kicks the power cord out of the wall (which happens). In this scenario,
a battery can be used to supply power temporarily. This can give the subsystem
enough power to finish the process of transferring the data from cache to disk safely.
Even backup batteries are often duplicated, just to be safe (see how crazy storage
100
The SAN Book III – October 22, 2001
people can be). As a magnetic medium, disk drives do not need continuous power to
retain data so once the data is transferred to disk, it is safe even if the battery
completely depletes or fails.
Bus Ctr
SCSI Bus
lr
Interf
Ctr
Logic Cache lr
FC Ctr
Interf lr
A subsystem with a bus architecture can increase its bandwidth by utilizing more
buses.
101
The SAN Book III – October 22, 2001
Bus Switch
102
The SAN Book III – October 22, 2001
Ctr
SCSI
lr
Interf
Switch Cache Switch
Ctr
Logic lr
FC Cache Ctr
Interf lr
103
The SAN Book III – October 22, 2001
NAS Subsystems
➤ NAS subsystems differ from traditional subsystems due to their
connectivity to the existing infrastructure (LAN) rather than an
emerging dedicated storage infrastructure (SAN).
NAS Configurations
In Figure 101 below are a couple of NAS configurations. In a traditional NAS
appliance, the server, RAID intelligence and drives are all within the same box. In
EMC’s Celerra, a general-purpose server is replaced with a file server (thin server
optimized for file serving), which is then attached to EMC’s Symmetrix subsystems.
Figure 101. Several NAS Configurations Are Possible
File File
RAID Drive RAID Drive
Server Server
Source: Salomon Smith Barney
High-end NAS
This is the original NAS market, pioneered by Auspex and now dominated by
Network Appliance. According to Dataquest, the high-end segment constituted the
largest revenue segment in 2000.
According to Dataquest, According to Dataquest, high-end or enterprise NAS devices are server appliances
high-end NAS devices for high-performance, high-capacity file sharing either as a resource for use by
are server appliances for
high-performance, high-
networked clients or within a SAN. They typically have RAID support, often have
capacity file sharing multiple storage controllers, and support multiple networking topologies such as
either as a resource for 100Base-T, Fibre Channel, FDDI, and ATM. Pricing is typically over $100,000
use by networked clients with capacities scaling to 6 TB.
or within a SAN.
Other definitions suggest that high-end NAS appliances include connectivity to
several network file systems, sometimes extending to proprietary systems and
mainframes, and performance is improved through automatic load and path
balancing and high-speed transfers between storage domains assigned to different
file or application servers. Accessibility is enhanced by a feature that ensures
disaster protection referred to as High Availability (HA). Failing components are
diagnosed, reported to a remote service site, and eventually corrected or replaced
104
The SAN Book III – October 22, 2001
105
The SAN Book III – October 22, 2001
Source: EMC
Celerra is a unique Celerra is a unique approach to NAS in that it is not a complete NAS appliance. It is
approach to NAS in that a series of NAS servers (thin file servers) without the traditional disk integration.
it is not a complete NAS
appliance. It is a series
Celerra attaches a Symmetrix disk subsystem (EMC’s traditional disk subsystem) on
of NAS servers without its back end, potentially forming a type of SAN architecture behind Celerra (we
the traditional disk would refer to this design as a NAS/SAN hybrid). In other words, Celerra provides
integration. a gateway for clients to directly access files over the LAN (typically Symmetrix disk
subsystems are directly attached to application servers or SANs, thereby forcing
clients to traverse application servers in order to retrieve files). Figure 104 below
illustrates how a Celerra-based NAS architecture might look.
106
The SAN Book III – October 22, 2001
Key Considerations
1)Relatively low cost of
Client 1
Server 1
management
Symmetrix 1
2)Fault tolerant
Client 2
3)Data transmission speeds
LAN Celerra 1
are somewhat constrained
by the LAN
Symmetrix 2
4)Heterogeneous
Client 3
connectivity
5)Scalable at the cost of
losing its simplicity
Client 4 Celerra 2 Symmetrix 3
Midrange NAS
Mid-range NAS devices According to Dataquest, midrange NAS devices typically have SCSI drives
support a number of (although ATA ones are starting to gain traction), a mixture of software and
drives and RAID, and
offer RAS (reliability,
hardware RAID controllers, and more storage management software functionality
availability, and such as clustering. Pricing ranges from $5,000 to $100,000 with capacities scaling
serviceability) features. to less than 2 TB. Other definitions suggest that these products typically have no
single point of failure and allow field-replaceable units to be swapped while the
system continues to operate (referred to as “hot swapping”). They incorporate some
performance features, such as faster write for RAID configurations, complex
caching algorithms, and/or dedicated data paths.
These systems can be scaled many ways, allowing easy growth in performance,
resiliency, and connectivity. Integrated backup is often offered as an option, and
administration of all network-connected storage can typically be performed from
one central location.
Examples of midrange NAS devices include Procom’s NetFORCE 1700, Network
Appliance’s F740, and Compaq’s TaskSmart N2400. In December, 2000, EMC
introduced its CLARiiON IP4700, a mid-tier NAS product (which can also scale to
the high end) with fault tolerant hardware, a wide range of network connections
including Gigabit Ethernet, and internal failover (processors, power supplies,
pathways). The IP4700 is available in a rack-mountable configuration and supports
from ten to 100 disks utilizing building block arrays. The array utilizes either 18- or
36-GB disks with a total capacity of up to 3.6 TB. EMC noted that the average
setup time has been ten minutes. The operating system is based on software from
CrosStor, which EMC acquired on November 1, 2000.
The Quantum Snap Server ES12 has 1 TB of raw capacity in a 3U rack-mountable
form factor with hot-swappable drives and hot-swappable, redundant power supplies
for under $25,000. Quantum recently acquired the Connex NAS business from
Western Digital to add additional enterprise-class functionality to its product
107
The SAN Book III – October 22, 2001
offering which could move it up the value chain (to the high-end market). Figure
105 below pictures EMC’s CLARiiON IP4700, which can also scale to the high end.
Figure 105. EMC’s CLARiiON FC4700
Source: EMC
Common applications include additional capacity for small and remote offices as
well as file serving, desktop backup, software distribution, and temporary “scratch”
space for IS administrators in large organizations. Common departmental and
enterprise NAS applications include capacity expansion and high-capacity, high-
performance file sharing.
Entry NAS
Dataquest defines entry- Dataquest defines Entry NAS devices as having one to four ATA hard disk drives.
level NAS devices as The low-end of Entry will typically contain one or two drives with either no RAID
having one or two hard
disk drives and no
or RAID 1 while the high-end of Entry will have four drives with RAID 5 support.
expandability. Pricing is usually less than $5,000 with raw capacity up to 300 GB. Other
definitions suggest that these products may be resistant to the most common failures
(such as disk drive, power supply, and cooling system) but are still vulnerable to less
common component failures and usually must be turned off for planned
maintenance.
Connectivity is sometimes limited to one topology (Ethernet, for example).
Common applications include additional capacity for small office/home office, small
businesses, and remote or branch offices, as well as file serving, desktop backup,
software distribution, and temporary “scratch” space for IT administrators in large
organizations. There is much evangelizing that could be done in this nascent
market, as we believe much of the target customer base is not aware of these
products and the benefits they offer. Examples of low-end NAS devices include
Quantum’s Snap Server and Maxtor’s MaxAttach Desktop family.
Figure 106 below pictures Quantum’s Snap 4100.
108
The SAN Book III – October 22, 2001
Source: Quantum.
Maxtor’s product also In July, 2001, Maxtor announced general availability for its 400 GB, 1U high
has dual 10/100 Ethernet Maxtor MaxAttach NAS 4300 File Server. The MaxAttach 4300 is based on the
ports, gigabit Ethernet
(320 GB), and a SCSI
Windows operating system, and we believe positions the company strongly at the
port for local tape high end of the entry group. Maxtor’s product also has dual 10/100 Ethernet ports,
backup. gigabit Ethernet (320 GB), and a SCSI port for local tape backup. We note that
most NAS filers in this product category do not have Gigabit Ethernet connectivity
nor do they offer an external SCSI port for tape backup. In addition, Maxtor’s
MaxAttach NAS file servers are compatible with backup software from VERITAS,
Legato, and Computer Associates, several device-management tools and support
RAID 0, 1, and 5.
Figure 107 below features the MaxAttach 4100.
Figure 107. Maxtor’s MaxAttach 4100
10/100 Disk
Soft
Ethernet activity
power
monitor
109
The SAN Book III – October 22, 2001
110
The SAN Book III – October 22, 2001
Figure 108. A General Purpose Server Is More Complex — and Expensive — Than a NAS Appliance
111
The SAN Book III – October 22, 2001
RAID
Disk
OS
Disk
FS
Disk
VM
112
The SAN Book III – October 22, 2001
What Is a Block?
Disks are subdivided A block represents physical disk space. Blocks are often thought of as the lowest
into blocks, which are common denominator of storage (i.e., data). Disks are subdivided into blocks,
numbered and laid out
onto the round disk
which are numbered and laid out onto the round disk platters in concentric circles,
platters in concentric similar to the lines on a phonograph album. Blocks are grouped into tracks, sectors
circles, similar to the and cylinders, but this is beyond the scope of our report. Blocks are the lowest
lines on a phonograph
common denominator of data.
album.
Blocks are stable locations of physical disk storage. Blocks are written to (or filled
with data) in order. In other words, first block 0 is filled, then block 1, then block 2,
etc. This simplicity is part of the reason block storage is often preferred. With
simplicity comes greater reliability, better performance, lower CPU utilization and
an easy, common, standard unit of data.
However, using multiple blocks, as storage does, creates the need to intelligently
manage these blocks through grouping and organization (i.e., volume management
software).
Below in Figure 111, we illustrate the concept of block storage, which is physical
storage. One important thing to note is that multiple blocks can be stored on one
disk (each cylinder below represent one disk, multiple disks form a storage
subsystem). Further, block-based storage groups blocks together, one next to the
other, until a disk is full. This means that data is saved in the order it is created
(conceptually). Therefore, when you go to retrieve data (composed of multiple
113
The SAN Book III – October 22, 2001
blocks), with block-based storage they can all be next to each other (conceptually,
although not always), which would make them easy to find. In other words, it’s
easier to find a box of needles in a haystack, than a couple of hundred individual
needles scattered about (as is sometimes the case with file-based storage).
Figure 111. Block-Based Storage
B Blocks
A
B
C
Subsystem
Source: Salomon Smith Barney
Volume Management
Volume management is Volume management is the management of block data. While blocks are the lowest
the management of block common denominator of data storage, they can be organized into logical groups via
data.
volume management. In other words, volume management technology operates at
the volume level, a step above the block level. Think of it as similar to school. The
twelfth grade is not a person (a block), it’s a way to more logically arrange people
(data blocks) and manage them (in this case to teach) in a more unified and
organized way. Think of blocks as people and volumes as grades. Below in Figure
112 we depict blocks aggregating into volumes.
114
The SAN Book III – October 22, 2001
Connectivity
A B C
Subsystem
Source: Salomon Smith Barney
115
The SAN Book III – October 22, 2001
What Is a File?
Think of files as Files are collections of blocks, but are more abstract than physical blocks. Think of
abstractions of blocks. files as abstractions of blocks. Files are not bound to any physical location. Files
are given names and organized in a non-linear format. In other words, while block
zero, block one, and block two are typically laid out one after the other, file X, file
Y, and file Z could all be on different volumes and/or on separate disks and/or
storage subsystems.
Below in Figure 114, we have illustrated how files might be logically grouped and
stored.
116
The SAN Book III – October 22, 2001
Volumes
A File X
B File Y Files
File Z
C
A B C
Subsystem
File Systems
Since files are abstract, Since files are abstract, they create a more virtualized environment that can be more
they create a more flexible than a physical environment. Put another way, blocks can be renamed into
virtualized environment
that can be more flexible
files and grouped logically (the same concept applies to volumes). Consequently,
than a physical file system software creates abstractions of blocks to improve data management (as
environment. illustrated below in Figure 115). For example, file one might be blocks 17–22 and
file two could be blocks two, nine to 14, and 23. Thought of another way, blocks
one to four could be file 17 and block five could be file two.
117
The SAN Book III – October 22, 2001
Connectivity
Server
So why make things so confusing with file system software and logically, rather
than physically, group data? Imagine if you physically stored data A, B, and C as
we did in the block storage diagram (Figure 111). Then we deleted data B. Then we
stored data D, E, and F. According to block storage layouts, block D would be laid
out after block C, leaving the space allocated to block B empty. This creates “holes”
of empty disk space which must then be logically filled with data. Unlike a file
system, a volume of block storage has no notion of “free,” “empty,” or “hole.” A
block is a block is a block, and all blocks exist on the volume continuously. One of
the things a file system adds is the notion of blocks being in use or free (a database
has a similar notion internal to the database). Bottom line, of course there is a
reason for the madness.
High performance file The better the volume management and file system technologies, the better the
system software can access to data. High performance file system software can reduce search times by
reduce search times.
providing quality data organization, including look ahead, and better utilize disk
capacity. This is similar to having a really good office filing system, rather than
simply tossing papers onto a pile and then sifting through the entire pile when a
single piece of data is needed. If it’s stored, organized, and managed optimally, it’s
easier to find and use. It takes less brain power to find it (processor utilization) and
it can be found faster (performance).
Think of file systems similar to Microsoft Explorer. Imagine you create three
directories (similar to volumes): A, B, and C. As employee data is organized it is
grouped into directories A , B, and C. Then Homer, an employee from group A
transfers to group B. Simply click on, drag and drop Homer’s data from directory
(volume) A to B. While the file layout has changed (i.e., Homer’s data is now
accessed under the B directory, instead of the A directory) the data (or blocks) can
remain in the same physical location.
118
The SAN Book III – October 22, 2001
119
The SAN Book III – October 22, 2001
Extent-Based Allocation
This method takes the Contiguous and Index methods a step further by combining
them and adding features to increase performance (the latter of which is its key
advantage) while decreasing disk capacity overhead and incorporating the use of
pointers (but cuts down on their use relative to the Linked and Index methods) and
disk capacity from holes.
Veritas and EMC’s Celerra use this method.
120
The SAN Book III – October 22, 2001
Tape
Library
Client 1
FS
Subsystem 1
Server 1
Client 2
LAN
FS Subsystem 2
Client 3 Server 2
Client 4 FS Subsystem 3
Server 3
Tape
Library
Client 1 FS
Server 1
Subsystem 1
Client 2
LAN SAN
FS
Client 3 Server 2
Subsystem 2
Client 4 FS
Server 3 Subsystem 3
121
The SAN Book III – October 22, 2001
Client 1
Server 1
Client 2
LAN
Client 3 Server 2
FS
Client 4 NAS
FS Appliance 1
NAS
Appliance 2
Architecturally
On the other hand, DAS Architecturally we can look at NAS as connecting to or obviating the need for
or SANs use simpler, application server file systems, which can reduce file system expenses (each copy of
more efficient volume
management (block)
a file system costs money) and more expensive application server CPU utilization
based storage as an (although NAS creates greater overall CPU demand, it offloads the CPU cycles from
interconnect. more expensive application servers onto less expensive file servers or NAS). On the
other hand, DAS or SANs use simpler, more efficient volume management (block)
based storage as an interconnect which is often viewed as more reliable and faster,
particularly for compute of data intensive applications. We have illustrated this in
Figure 119 below.
122
The SAN Book III – October 22, 2001
Application
Server
Application Software
NAS Operating System
File Server
Functionality File System
File System
Volume Management
Volume Management
RAID
Disk
Volume Management
RAID
Disk
Storage
Subsystem
Source: Salomon Smith Barney
Storage subsystems can Note: Storage subsystems can have their own volume management technology.
have their own volume Although it is uncommon, this has been one of EMC’s key differentiators on its high-
management technology.
Although it is
end subsystems.
uncommon, this has
been one of EMC’s key
Network Appliance’s File System
differentiators on its One of Network Appliance’s most distinctive advantages over traditional server-
high-end subsystems. attached storage, as well as its NAS competition, is its sophisticated file system,
which is strategically located on its NAS appliance’s thin server. In conventional
local or direct attached storage and block-level SAN implementations, the file
system is colocated with the operating system/application server — separate from
the RAID controller. In this approach, file systems send blocks to the RAID
controller, which lays the data blocks out on the disk array. By separating the file
system from the controller/subsystems, Network Appliance believes that file system
advantages such as security, logical volume and file management, and data
management can be lost (although many would argue the opposite).
Network Appliance’s filers integrate its file system — called Write Anywhere File
Layout (WAFL) — directly with the RAID array, which preserves and enhances
file-mode access advantages. Network Appliance believes benefits for local file-
sharing access include scalability, management, resiliency, failover, and lower total
cost of ownership. To date, Network Appliance filers have been attached to
application servers and clients through TCP/IP utilizing NFS, CIFS, and HTTP
protocols.
123
The SAN Book III – October 22, 2001
Heterogeneous Connectivity
In traditional server-attached storage architectures, storage subsystems typically
cannot share files between platforms. This is partly due to the location of the file
system, which is located on the application server. In effect, the location of a file
system in a traditional storage architecture couples each storage subsystem with the
operating or file system being used by the application server, as illustrated below in
Figure 120.
Figure 120. Storage Subsystem Connectivity by Operating and File System
Mainframe Mainframe
FS
UNIX UNIX
FS
NT NT
FS
Source: Salomon Smith Barney
Being able to provide Being able to provide heterogeneous connectivity between discrepant systems and
heterogeneous technologies is a powerful advantage. For example, would it be nice to store UNIX
connectivity between
discrepant systems and
and NT data onto a single storage subsystem? Network Appliance can due to its file
technologies is a system architecture, but so can EMC using its block-based architecture.
powerful advantage.
Heterogeneous data storage has been one of EMC’s largest competitive advantages
in the past. Recently many other storage subsystem vendors have developed
heterogeneous connectivity, although EMC still supports the broadest number and
versions of operating systems, including the mainframe. Block based storage
subsystems can store UNIX and NT files on the same subsystem and then translate,
if requested, NT data into UNIX data. However, due to its file system architecture
Network Appliance has taken heterogeneous connectivity to a more advanced level.
Block-based subsystems partition disks by operating system (e.g., NT, UNIX). It
can then cache the data and translate it to communicate with other operating
systems. In Figure 121 below, we have illustrated how a storage subsystem stores
data and retrieves it by operating system.
124
The SAN Book III – October 22, 2001
Mainframe
FS Subsystem
Mainframe
UNIX
FS UNIX
NT
NT
FS
What sets Network What sets Network Appliance’s heterogeneous storage apart from others is that it
Appliance’s actually stores data in a neutral (neither NT nor UNIX) format. This obviates the
heterogeneous storage
apart from others is that
need to “translate” data back and forth, using expensive caching buffers. However,
it actually stores data in it does add overhead at the onset of a save or write command. Below in Figure 122,
a neutral (neither NT nor we have anecdotally illustrated how Network Appliance stores data in a neutral
UNIX) format. format.
Figure 122. Network Appliance Uniquely Stores Data in a Heterogeneous, Neutral Format
A Neutral A A’
NT OR
NT Neutral
B
B B’
About 40% of all Network Appliance systems shipped are installed out of the box to
interconnect multiple computing platforms, typically UNIX and NT. By leveraging
its file-system design, Network Appliance is able to offer heterogeneous
connectivity while maintaining only one file. To offer heterogeneous connectivity,
125
The SAN Book III – October 22, 2001
other storage systems need to maintain multiple copies of the same file, increasing
complexity and requiring additional storage space.
Below in Figure 123 we have illustrated how Network Appliance’s heterogeneous
connectivity might be viewed. Note: Network Appliance does not fully support the
mainframe.
Figure 123. Network Appliance’s Heterogeneous Storage
NetApp
Appliance
UNIX
FS
UNIX &
NT
NT
Scalability
Simply snap new disk Files organize data logically, instead of physically (which is done by volume
capacity (i.e., physical mangers). This means that blocks can be any size, regardless of physical limitations.
disk) onto the network
and an intelligent file
Thought of another way, simply snap new disk capacity (i.e. physical disk) onto the
system can seamlessly network and an intelligent file system can seamlessly assimilate that new storage.
assimilate that new With volume management software, a new node must be initialized, which
storage. sometimes takes the environment temporarily offline. Blocks tend to be a fixed size
while files and file systems can be any size.
No Free Lunches
File-based storage also has its downfalls. For example, it requires higher CPU
utilization and is more complex than block-based storage. These complexities can
increase the chances of failure and negatively impact performance. Often times in
computing, simpler is better.
126
The SAN Book III – October 22, 2001
Tape
Library
Server 1
Client 1
Subsystem 1
Client 2
LAN Server 2
SAN
Client 3 Subsystem 2
Server 3
Client 4
Subsystem 3
FS
127
The SAN Book III – October 22, 2001
Server 1
Client 1
Client 2 Subsystem 1
LAN Server 2
Client 3
NAS File
Client 4 Server 1
Subsystem 2
Server 1
Client 1
Client 2 Subsystem 1
LAN Server 2
Client 3
NAS File
Client 4 Server 1
Subsystem 2
128
The SAN Book III – October 22, 2001
Figure 127. EMC’s HighRoads Can Leverage SANs for Best-in-Class Performance
Tape
Library
Server 1
Client 1
Subsystem 1
Client 2
LAN Server 2
SAN
Client 3 Subsystem 2
Server 3
Client 4
Subsystem 3
FS
Competition...Competition!!!
Its no wonder to us that Network Appliance, EMC, and Veritas are becoming fierce
competitors for data management and file management technology.
Virtual Reality?
Essentially, virtualization Virtualization is the natural evolution in data management. We started with disks
is a hierarchical (blocks), then grouped them into subsystems (volumes), and then organized the data
grouping of volumes that
allow the management of
in the server (file systems). Virtualization is the next logical evolutionary (not
all storage as a common revolutionary) step in data management. Essentially, it’s a hierarchical grouping of
pool. volumes that allows the management of all storage as a common pool. Below in
Figures 128 and 129 we have depicted how this might fit into our block diagram.
129
The SAN Book III – October 22, 2001
Connectivity
130
The SAN Book III – October 22, 2001
ubiquitous data management in the near term, we do not expect to see it in a broadly
usable form for a number of years.
Virtualization introduces Virtual Promise: Unlimited storage capacity on demand. In other words, capacity
the concept of “free” can be added on the fly without taking systems down or having to reconfigure them,
versus “in use” space to
the SAN at the
similar to the way file systems operate. Virtualization introduces the concept of
block/volume level and “free” versus “in use” space to the SAN at the block/volume level and the ability to
the ability to map things map things at a logical level all the way down to a physical level.
at a logical level all the
way down to a physical Below in Figure 130, we have illustrated the way a virtual pool of storage might be
level. interpreted graphically.
Figure 130. Virtual Storage Pool of Unlimited Storage on Demand
Virtual Storage
Pool
Analogy
Similar to our previous school system analogy, virtualization becomes another layer
or subset.
➤ person = block (disk)
➤ grade = volume (volume management)
➤ school division (primary school, high school, college, etc.) = virtualization
➤ school = file system
131
The SAN Book III – October 22, 2001
132
The SAN Book III – October 22, 2001
Software
133
The SAN Book III – October 22, 2001
134
The SAN Book III – October 22, 2001
135
The SAN Book III – October 22, 2001
Figure 131. Core Storage Management and Virtualization Software, UNIX platform
1999 2000
Other
Other 9%
EMC EMC
10%
23% 30%
Veritas
Veritas 61%
67%
136
The SAN Book III – October 22, 2001
137
The SAN Book III – October 22, 2001
138
The SAN Book III – October 22, 2001
139
The SAN Book III – October 22, 2001
12-Apr-95 10-Apr-95 Tidalwave Technologies Pooling of interests; 117,374 shares of VRTS; FirstWatch failover
Inc. exchange of stock VERITAS assumes a warrant software; HA
which entitles the holder to acquire
an additional 27,010 shares of
VRTS common stock
The advantages of a The advantages of a “total solutions” product offering have helped drive
“total solutions” product consolidation in the storage software market. Rather than offering a niche set of
offering have helped
drive the consolidation
software products and competing with larger competitors with a complete software
in the storage software product portfolio, both acquirers and acquirees see the wisdom in merging, thereby
market. increasing the combined companies’ potential addressable market by offering a
more complete total solutions package.
We estimate that 25%– Even with all the consolidation that has already occurred, we believe the storage
30% of the storage software market still has many companies left that will go out of business or will be
software market remains
“white box.”
acquired. We estimate that 15%–20% of the storage software market remains
“white box” (i.e., in the hands of small companies).
140
The SAN Book III – October 22, 2001
but will increase the value of the other more traditional storage software product
areas.
It has been the recent For example, the development and implementation of SRM and SAN/Device
momentum of data Administration (which are two pieces of the same function; please refer to the
replication, SAN/Device
Administration and SRM
“Software Applications” section for more details on SRM) have enabled
software that have backup/recovery networking to be done more manageably. When systems go on a
brought mainframe and network they often become more complex and harder to manage due to the inclusion
open-systems computing
of many more variables. With SRM and SAN/Device Administration, the
to a more level playing
field. environment can be viewed and altered more easily. SRM will bring the
architecture up into a format that is easy to read and understand and SAN/Device
Administration will alter the environment to make the appropriate policy changes
desired. Therefore, SRM and SAN/Device Administration help generate more
demand for Backup/Recovery Networking (open systems) versus Backup/Recovery
Host (mainframe).
Figure 140 below outlines the storage software market segments in 1998 and 2005
based on Dataquest’s projections and categories. Going back to 1998 might seem
like a bit much, but it helps accentuate the trends. Note the increase in the size of
the market: from $2.9 billion (1998) to $16.7 billion (2005) in seven years,
representing a 29% CAGR between 1998 and 2005. Also note that SRM moves to
10% from 4% of the total market; Virtualization to 7% from 2%; and Replication to
20% from 15%. Note: Dataquest revenue calculations include new license sales but
not revenues from maintenance and support services.
Figure 140. Storage Software Trends
1998 2005
$2.9 Billion $16.7 Billion
141
The SAN Book III – October 22, 2001
Most companies are It comes as no surprise to us that most companies are increasing their focus on NT
increasing their focus on and UNIX. In fact, while BMC, Computer Associates, and IBM derived a good
NT and UNIX.
portion of their revenues from the mainframe platform in 2000 (63%, 46%, and
76%, respectively), the up-and-comers such as EMC and Veritas derive the bulk of
their revenues from NT (23% and 32%, respectively) and UNIX (64% and 64%,
respectively). The mainframe storage software market share breakdown for 2000 is
shown in Figure 142 below.
Figure 142. Mainframe Storage Software Market Share (2000)
Other
StorageTek 10%
9%
IBM
BMC
42%
9%
EMC
11%
Computer
Associates
19%
Figure 143 below illustrates the storage software market share on the UNIX
platform in 2000. Veritas has a much larger presence in open systems, on both
UNIX and Windows NT.
142
The SAN Book III – October 22, 2001
IBM
7%
Veritas
26%
Figure 144 below illustrates the storage software market share on the Windows NT
platform in 2000.
Figure 144. NT Storage Software Market Share (2000)
Other EMC
22% 26%
Legato
6%
Compaq Veritas
8% Computer 22%
Associates
16%
143
The SAN Book III – October 22, 2001
2000 2005
Market Size Market Size
Total $5,251M Total $16,729M
CAGR 26.1%
Storage
Resource
$271.3 Mgmt. $1,709.5; 44.5%
NA Clustering NA
144
The SAN Book III – October 22, 2001
Figure 146. Total Worldwide Storage Software Market Revenues by Selected Vendor
($ in millions)
1998 1999 2000
Market Market Market Change
1998 Share 1999 Share 2000 Share 1999-2000
EMC 444.7 15% 802.8 20% 1340.6 26% 67.0%
Veritas 293.2 10% 504.7 12% 855.1 16% 69.4%
IBM 607.7 21% 723.6 18% 843.9 16% 16.6%
Computer Associates 596.2 21% 798.7 20% 612.7 12% (23.3)%
BMC Software 164.9 6% 228.9 6% 224.1 4% (2.1)%
Network Appliance 14.9 1% 53.1 1% 156.0 3% 193.8%
StorageTek 86.9 3% 112.8 3% 143.5 3% 27.2%
Compaq 14.8 1% 44.9 1% 143.2 3% 218.9%
Legato 147.2 5% 172.2 4% 142.3 3% (17.4)%
Source: Dataquest (April, 2001)
145
The SAN Book III – October 22, 2001
146
The SAN Book III – October 22, 2001
fs
Software Applications
➤ The four storage software applications we view as important
catalysts for future growth are: 1) Storage Resource Management, 2)
Replication, 3) Serverless and LANless Backup, and 4) Hierarchical
Storage Management. [We discuss file systems, volume
management, and virtualization in our previous section: “RAID Is No
Good Without Core Software.”]
➤ We expect to see these key four segments also drive growth in more
traditional storage software product areas.
➤ We believe EMC and Veritas are the two horses to bet on. Also, there
are other emerging companies coming out with new applications and
technologies that deserve attention.
The four storage software applications we view as important catalysts for future
growth are: 1) Storage Resource Management, 2) Replication, 3) Serverless and
LANless Backup, and 4) Hierarchical Storage Management. [We discuss file
systems, volume management, and virtualization in the previous “RAID Is No Good
Without Core Software” section.] We expect to see these key four segments also
drive growth in more traditional storage software product areas.
We believe EMC and Veritas are the two horses to bet on in the storage software
race. They are currently the storage software market leaders and both gained market
share in 2000. We also believe other companies are coming out with important new
products and technologies that merit attention.
147
The SAN Book III – October 22, 2001
What Is SRM?
While SRM has traditionally been the software that provided a window to view the
storage environment from a single console and measure the performance of devices,
it is also increasingly including the device administration software that provides the
administrative tools to alter the environment, including policy definitions, device
zoning, load balancing, and capacity allocation. In other words, SRM used to be
defined as the software that made everything easy to view and measure. Now it
includes active management, referred to as “drag-and-drop” functionality. Think of
it in the context of your Microsoft Explorer application, separating viewing the
environment from using your mouse to make a change to it.
EMC’s ControlCenter 4.3 has been the clear leader in the Storage Resource
Management software market. We expect some new entrants to gain traction in
2002 and beyond, for example, Prisa’s VisualSAN, Veritas’s SANPoint Control 2.0,
InterSAN’s Virtual Private DataPath technology, Legato’s GEMS SRM, and Sun
Microsystems’ HighGround SRM 5.0.
Below in Figure 147 is a snapshot of EMC’s ControlCenter.
Figure 147. EMC’s ControlCenter Software
Symmetrix Manager
Device Administration
Device Administration software controls a device; in other words, it can change its
environment — it is active, not passive. Brocade’s Fabric OS software is an
excellent example of device administration software. Another example is EMC’s
Symmetrix Manager, which is now tied together with ControlCenter.
148
The SAN Book III – October 22, 2001
Load Balancing
Brocade’s Fabric OS is Brocade’s Fabric OS is considered to be one of the more sophisticated storage
one of the more networking software device administration software applications. It helps create
sophisticated storage
networking software
network-aware storage environments. Features such as load balancing ensure the
device administration even flow of data through the network. This further contributes to data availability
software applications. by balancing traffic evenly across the network. In other words, if one switch (or
port) is being overloaded and another switch (or port) has excess capacity, some of
the data can be sent through the other, less congested switch (or port), resulting in a
more efficiently run network.
Trunking is also a component of load balancing. Trunking allows the aggregation of
the bandwidth from multiple ports. For example, four 2 Gpbs ports can be trunked
together to create a single virtual 8 Gbps port.
Failover
Another one of Brocade’s advanced device administration software features has
been its failover capabilities. This is the software that enables multiple networking
devices to reroute storage traffic through alternative paths upon the devices’ failure
(failover can also occur within the same device between multiple ports upon
component failure).
Security
Security has been increasing in importance to protect users from accessing data
without permission and protecting against corruptions and failures.
We believe Brocade’s advanced software features have been one of its strongest
assets that have helped it dominate the storage switch market.
149
The SAN Book III – October 22, 2001
150
The SAN Book III – October 22, 2001
Figure 149 below has more examples of Prisa Networks’ GUI interface.
Figure 149. Prisa Networks’ VisualSAN Performance Manager
$1,800
$1,600 44.5% CAGR
$1,400
$1,200
$1,000
$800
$600
$400
$200
$0
1998 1999 2000 2001 2002 2003 2004 2005
Source: Dataquest (May, 2001). CAGR is for 2000 through 2005.
151
The SAN Book III – October 22, 2001
While vendors such as Computer Associates, BMC Software, and IBM represent
mainframe SRM software and currently own the bulk of the market, the entrances of
Sterling (owned by Computer Associates), HighGround (owned by Sun), and EMC
(through its ControlCenter software) have pushed SRM into open systems and are a
main driver for Dataquest’s projections (see Figure 150).
Figure 151 below breaks out the market share of industry leaders on all platforms.
Figure 151. Storage Resource Management Market Share, All OS
1999 2000
HighGround (Sun) Others HighGround (Sun) Others
EMC
2% 17% 3% 14%
19% EMC
IBM IBM 33%
7% 7%
Figure 152 below breaks out the market share of industry leaders on the UNIX
platform. EMC has the largest share of this market.
Figure 152. Storage Resource Management Market Share, UNIX
1999 2000
Hewlett-Packard Legato
2% Others
Others 1%
EMC 8%
18% HighGround (Sun)
31% 2%
Legato EMC
BMC Software
17% 54%
33%
Hewlett-Packard BMC Software
4% 27%
HighGround (Sun)
3%
Figure 153 below breaks out the market share of industry leaders on the Windows
NT platform. EMC gained significant share from 1999 to 2000.
152
The SAN Book III – October 22, 2001
1999 2000
EMC Others
Others EMC
2% 23%
21% 31%
153
The SAN Book III – October 22, 2001
2. Replication
Replication is We view replication as a critical business asset. Replication creates multiple copies
transitioning from of data to ensure an uninterrupted business environment. A large part of replication
becoming a
discretionary, to a
is that it acts as an emergency plan for data infrastructures that can prevent a
necessary, expense. temporary or permanent loss in data availability. Therefore, replication is
transitioning from becoming a discretionary, to a necessary expense and in some
cases is being leveraged as a competitive advantage as data grows both in quantity
and significance and customers demand access to information 24 hours a day, seven
days a week without interruption for any reason.
Replication has been a solid area of growth for EMC in the past and we think it’s
only going to get better.
What Is Replication?
Replication enables Replication enables the mirroring of data between disk storage subsystems to ensure
mirroring of data the continuous availability of enterprise data during planned or unplanned outages.
between two physically
and geographically
Replication provides business continuance capability in the event of a data center
separated storage disaster, and during planned events such as daily backups, database loads and
subsystems. refreshes, application testing, scheduled maintenance, and data center migrations or
consolidations.
Below we have illustrated what data replication might look like.
Figure 154. Data Replication
Primary Data
Replicated Copy
154
The SAN Book III – October 22, 2001
Although Veritas recently introduced its own replication solution, EMC is still the
clear leader in this market segment. Other players include Legato’s Octopus,
NetApp’s SnapMirror, and NSI’s Double-Take.
Disaster Recovery (DR)
Disaster Recovery software, a subset of Replication, enables the high availability
(HA) of data by creating a real-time copy at a remote location. In the event of a
disaster, such as a power outage, which could affect the data infrastructure in a
particular business or region, businesses can be switched over to a replicated data
facility in a different location.
Without replication, lost data and prolonged downtime could result in a loss of
massive amounts of revenue and productivity, as well as in customer trust, brand
equity, and competitive advantages which can take years to build but just hours to
destroy. Replication can also prevent a loss in productivity as employees sit idle for
hours while the original data infrastructure is restored, as would be the case with a
recovery from a tape backup.
Below in Figure 155, we illustrate how a disaster recovery architecture might be
viewed.
Figure 155. Disaster Recovery Architecture
Primary data
center in NY
Backup data
center in FL
155
The SAN Book III – October 22, 2001
$3,500
$3,000 28.0% CAGR
$2,500
$2,000
$1,500
$1,000
$500
$0
1998 1999 2000 2001 2002 2003 2004 2005
Source: Dataquest (May, 2001). CAGR is for 2000 through 2005.
Data Replication has been one of EMC’s most valuable assets, generating 42% of
EMC’s new software license revenues in 2000. According to Dataquest, EMC’s
replication software has dominated its segment with 58% market share in 2000,
while its three closest competitors combined represented only 19% (see Figure 158
below). According to Dataquest, EMC grew its replication revenues 38% in 2000.
Note: EMC replication revenues include SRDF, TimeFinder, InfoMover,
CopyPoint, and Extractor/Resource Pak.
156
The SAN Book III – October 22, 2001
1999 2000
Others Others
StorageTek
19% 19%
4%
StorageTek
5% Symantec
4%
Symantec Compaq
6% EMC
EMC 5% 58%
Compaq 65%
Network
1% Appliance
Network 10%
Appliance
4%
Figure 159 below illustrates the UNIX replication market share. EMC clearly
dominates this market segment.
Figure 159. Replication Market Share, UNIX
1999 2000
Others Others
11% 15%
EMC
EMC
85%
89%
Figure 160 illustrates the Windows NT replication market share. EMC has
significantly increased its lead in this segment.
157
The SAN Book III – October 22, 2001
1999 2000
Others Others
21% EMC 15%
29% Legato EMC
9% 41%
PowerQuest
Legato
11%
21% Compaq
PowerQuest 13% Compaq
16% 24%
SRDF Characteristics
➤ Storage-centric: SRDF provides this facility simultaneously to mainframe,
UNIX, Windows NT, and other systems. This storage-centric (i.e., residing on
the subsystem), mirrored, data storage solution duplicates production site data
on one or more physically separate target Symmetrix systems. Systems can be
across the room, across the globe, or anywhere in between.
➤ Low Transmission Costs: SRDF’s multi-hop capability decreases line
transmission costs by mirroring only the tracks changed since the last update,
and by taking advantage of lower cost transmission lines, such as Ethernet.
➤ Simultaneous Use of Data: Scheduled backup operations, data warehouse
loading, or new application testing can be performed from the target Symmetrix
data (the second copy), while normal operations are continuously running on the
original Symmetrix system.
➤ Synchronous and Asynchronous: EMC mainly offers SRDF over Fibre
Channel, interconnected by T1 and T3 lines, for synchronous applications.
EMC also offers SRDF over IP (Internet Protocol) for asynchronous
applications as a less expensive, less robust solution. (Note: Synchronous is real
time, asynchronous is for static information.)
Below is an illustration of a potential architecture using EMC’s SRDF. In one
branch, two Symmetrix systems are connected via dark fiber. In this configuration,
replication can be accomplished either synchronously or asynchronously. In another
branch, two Symmetrixes are attached via IP. In this configuration, replication can
only be accomplished asynchronously.
158
The SAN Book III – October 22, 2001
Figure 161. EMC’s Replication Software Can Mirror Data in Two Different Locations
Tape
Library
EMC
Client 2
LAN SAN Symmetrix 2
159
The SAN Book III – October 22, 2001
160
The SAN Book III – October 22, 2001
Tape
Library
DAS
(Traditional Storage Backup)
Client 1
Server 1 •In a traditional DAS
architecture, backup traffic
Subsystem 1
must traverse an already
Client 2
LAN congested LAN (often at
night or during non-business
Server 2
hours).
Client 3 Subsystem 2
•DAS backups involve
application servers to run the
backup application software.
Client 4
Server 3
Subsystem 3
161
The SAN Book III – October 22, 2001
server involvement (see Figure 164). Therefore, serverless backups are also
LANless by definition.
Figure 164. SAN Backups Can Be LANless and Serverless
Tape
Library
SAN
Client 1 (Serverless Backup)
Server 1
•In a SAN architecture,
backup traffic can be
Subsystem 1 offloaded onto a dedicated
Client 2
LAN SAN storage network to
avoid further LAN
Server 2 congestion.
Client 3
Subsystem 2
•Backup and recovery do
not have to involve an
application server to run the
Client 4
Server 3 backup application software,
Subsystem 3 thereby freeing server
capacity.
162
The SAN Book III – October 22, 2001
HSM Negatives:
➤ Slower data access
SANs enable a broader application of HSM, even though HSM has been around for
a while. SANs make HSM more attractive by reducing the resources needed to
utilize it.
HSM is a good example of how the adoption of a SAN infrastructure can make
traditional applications even more powerful. Additionally, as IT budgets come
under more scrutiny, applications with immediate and tangible cost savings should
grow in importance. Therefore, we believe HSM is important to understand.
What Is HSM?
Hierarchical Storage Hierarchical Storage Management is just what it sounds like: prioritized data
Management (HSM) is storage. HSM assigns a level of “usage” to data. By doing so, intelligent HSM
just what it sounds like:
prioritized data storage.
software can prioritize the accessibility of data and store it according to
predetermined parameters. For example, HSM would locate a frequently used file
(say a phone list) where it could be readily accessed on a primary storage device.
On the other hand, HSM might migrate less frequently used data (with
predetermined and preset parameters and thresholds; for example, an employment
contract) to a secondary storage device. In general, HSM is designed to migrate less
frequently used data from primary, quick-access storage to secondary, less
expensive, slower-access storage. Below are definitions of primary and secondary
storage:
Primary storage
A primary storage device is a storage device or subsystem that provides fast,
easy access to data, such as a disk subsystem.
Secondary storage
A secondary storage device is a storage device or subsystem that provides
less expensive, slower access to data, such as a tape library or optical
jukebox.
In Figure 165 below, we have outlined how HSM might view data.
163
The SAN Book III – October 22, 2001
The way
HSM Less Frequently
Looks At It Used Data
Data
Frequently
Used Data
Source: Salomon Smith Barney
The goal is to reduce the HSM can migrate data based on many different variables, or by combining
cost of storing data by numerous variables, such as frequency of use, date (i.e., file age), file size, and type
more effectively utilizing of data (e.g., attachments, graphics). An example might be email attachments and
secondary storage
capacity. Secondary
emails two weeks old and older. In this example, all attachments and emails that are
storage can cost as little two weeks old or older would be migrated from a disk storage subsystem to a tape
as 3% of the cost of library. The goal is to reduce the cost of storing data by more effectively utilizing
primary storage.
secondary storage capacity.
OTG Software provides a The cost advantage of HSM does not come without a performance penalty. Here
unique indexing scheme again, we run into the importance of data access, rather than data storage. This
that truncates (migrates)
functionality can be a key differentiator. For example, OTG Software provides a
target data from its
location on primary unique indexing scheme which truncates (migrates) target data from its location on
storage, leaving behind a primary storage, leaving behind a small “stub file.” The stub file has the updated
small “stub file.” location of the target data. When clients access data that has been migrated, they hit
the stub file, which, in turn, accesses the target data directly from secondary storage
without first having to copy it back to primary storage. In Figure 166 below, we
have illustrated how HSM effectively maximizes storage capacity.
164
The SAN Book III – October 22, 2001
Figure 166. HSM Data Migration — Freeing Primary Storage Capacity by Utilizing Less Expensive
Storage
Tape Library
Disk Subsystem Disk Subsystem
Using HSM
Unused Capacity
Data
Less Frequently Migration
Used Data Stub Files HSM Migrated Data
165
The SAN Book III – October 22, 2001
Tape
Library
DAS
(Traditional Storage Backup)
Client 1
Server 1 •In a traditional DAS
architecture, backup traffic
Subsystem 1
must traverse an already
Client 2
LAN congested LAN (often at
night or during non-business
Server 2
hours).
Client 3 Subsystem 2
•DAS backups involve
application servers to run the
backup application software.
Client 4
Server 3
Subsystem 3
Below, in Figure 168 we illustrate the same data migration as above, but with a SAN
architecture.
Figure 168. Data Migration in a SAN Architecture
Tape
Library
SAN
Client 1 (Serverless Backup)
Server 1 •In a SAN architecture,
backup traffic can be
Subsystem 1 offloaded onto a dedicated
Client 2
LAN SAN storage network which
avoids further LAN
Server 2 congestion.
Client 3
Subsystem 2
•Backup and recovery does
not have to involve an
application server to run the
Client 4
Server 3 backup application software,
Subsystem 3 thereby freeing server
capacity.
Using a SAN, no servers or LAN need to be accessed for tape backup, increasing the
availability of these resources for other uses.
166
The SAN Book III – October 22, 2001
$700
$600 15% CAGR
$500
$400
$300
$200
$100
$0
1998 1999 2000 2001 2002 2003 2004 2005
Source: Dataquest (May, 2001). CAGR is for 2000 through 2005.
Figure 170 below shows HSM and Archive market share. IBM dominates this
market segment.
Figure 170. HSM and Archive Market Share, All OS
1999 2000
Others Others
25% 27%
IBM IBM
49% 48%
SGI SGI
12% 8%
Figure 171 below shows HSM and Archive market share on the UNIX platform.
The market is fairly fragmented with no one having more than 16% market share.
167
The SAN Book III – October 22, 2001
1999 2000
FileTek
FileTek
15%
17% Others
Others
36% 37% SGI
SGI
14%
12%
Veritas Legato
IBM IBM
3% 3%
11% Sony ADIC 12%
Legato ADIC
Veritas 5% 9%
Sony 8% 10%
5%
3%
Figure 172 below illustrates HSM and Archive market share on the Windows NT
platform. Computer Associates and OTG Software are the leaders in this market.
Figure 172. HSM and Archive Market Share, Windows NT
1999 2000
Others
Others OTG Software 20%
22% 25%
KOM Networks OTG Software
4% 42%
Smart Storage
KOM Networks
13%
3%
Computer Computer
Smart Storage
Associates Associates
11%
39% 21%
168
The SAN Book III – October 22, 2001
169
The SAN Book III – October 22, 2001
170
The SAN Book III – October 22, 2001
Networking
Storage Networking
Storage Networking I/O
Storage Networking Equipment
171
The SAN Book III – October 22, 2001
Storage Networking
➤ We believe storage networking is on the cusp of resuming solid
growth in 2002, due to many product launches, such as Brocade’s
Silkworm 12000, and the industry’s migration to 2 Gbps and PCI-X.
Disk 1
Server 1
Disk 2
Server 2
Tape 1
Server 3 Disk 3
LAN
Disk 4
Server 4
Disk 5
Server 5 Disk 6
Server 6 Disk 7
Disk 6
NAS
Server 1 Disk 7
NAS
Server 2 Tape 2
Disk 8
Note: In the diagram we use switches, which are usually purchased with 16 ports
(moving to include 32 ports). We could use Directors instead or as well. Directors
172
The SAN Book III – October 22, 2001
come with 64 ports (McData) and 128 ports (Inrange). We do not specifically depict
any routers or hubs per se. Also, note the redundancies in order to ensure multiple
paths and greater uptime in case of a path or device failure.
173
The SAN Book III – October 22, 2001
Storage
Mainframe ESCON
Directors
Market Breakdown
SAN is one of the most We believe the SAN concept is one of the more exciting killer applications emerging
intriguing killer in today’s information-centric world. But what actually goes on inside the SAN
applications to emerge.
cloud?
We have positioned the various storage networking components in Figure 175
below, with the upper portion of the pyramid representing more
sophisticated/higher-cost devices. We have also depicted the segments we believe
are current growth areas. Figure 175 essentially represents our view of the storage
networking market.
174
The SAN Book III – October 22, 2001
SAN
APPLIANCE
DIRECTOR
Technology
SWITCH
Cost
R
TE
(FABRIC)
U
RO
SWITCH (FC-AL)
BRIDGE
HUB
HBA
Note: The terms and names of storage networking equipment are very similar to
LAN and WAN networking equipment (i.e., bridges, routers, hubs, and switches) but
are, in fact, very different technologies.
Projections
Storage Networking Growth
Storage networking has emerged as one of the highest-growth areas of storage. We
expect storage networking growth to lead the overall storage industry. According to
Dataquest, storage networking growth (2000–05) is expected to reach a 67.0%
CAGR versus storage software’s estimated 26.1% CAGR and RAID-based disk
storage 14.7%.
175
The SAN Book III – October 22, 2001
Figure 176. Storage Networking Growth Is Expected to Lead the Industry (YoY Growth Projections)
250%
200%
150%
100%
50%
0%
1999 2000 2001 2002 2003
Projections
Dataquest projects storage networking revenues will grow to $16.9 billion in 2005
from $1.3 billion in 2000, representing a 67.0% CAGR. This projected high-growth
market has attracted fierce competition from many companies looking to be the
Cisco of storage — including Cisco. (See the “Fibre Channel Versus Ethernet”
section for more details).
Figure 177. Projected Storage Networking Revenue Growth
CAGR
(Revenue in $millions) 2000 2001 2002 2003 2004 2005 2000-2005
FC Routers $50 $70 $110 $160 $240 $360 48.4%
FC Hubs $50 $60 $60 $60 $60 $60 3.7%
FC Switches $470 $710 $1,090 $2,210 $4,090 $4,930 60.0%
FC Integrated Switches $0 $90 $180 $460 $840 $1,040 NM
FC HA Core Products $190 $400 $1,040 $2,390 $4,620 $6,240 101.0%
FC HBAs $540 $810 $1,140 $2,090 $3,480 $4,250 51.1%
Total Storage Forecast $1,300 $2,140 $3,620 $7,370 $13,330 $16,880 67.0%
Year-to-Year Growth -- 64.6% 69.2% 103.6% 80.9% 26.6% --
Source: Dataquest (May, 2001)
Platform Focus
Windows NT is clearly expected to be a SAN growth driver. This should bode well
for Brocade (which we believe has the lion’s share of the NT storage networking
market), Emulex (which supplies Fibre Channel HBAs to Compaq, and EMC), and
QLogic (which supplies Fibre Channel HBAs to Dell and NetApp).
176
The SAN Book III – October 22, 2001
Using Dataquest’s JOBD and external RAID Fibre Channel storage estimates and
projections as a reference point, we estimate that NT should grow from
approximately 30% of the SAN market in 2000 to approximately 57% in 2005.
Note: Although UNIX is projected to surrender market share to NT, what is not
captured below is the fact that NT is also growing nicely. In other words, NT is
simply expected to outpace UNIX growth. Also, since FICON has only recently been
released, it has not yet been included in this analysis.
Figure 178. SAN Market Share by Platform
2000 2005
Other
Other 1%
3%
NT
30%
UNIX
42%
NT
57%
UNIX
67%
177
The SAN Book III – October 22, 2001
$4,500
$4,000 51.1% CAGR
$3,500
$3,000
$2,500
$2,000
$1,500
$1,000
$500
$0
2000 2001 2002 2003 2004 2005
Source: Dataquest
178
The SAN Book III – October 22, 2001
What Is an HBA?
A Host Bus Adapter Much like a Network Interface Card (NIC) interconnects servers to PCs via the LAN
(HBA) interconnects or Ethernet, a Host Bus Adapter (HBA) interconnects servers to external storage
servers-to-external
storage devices.
devices (note: HBAs provide far more functionality than NICs, which we discuss in
the “Anatomy of HBAs” section). Back when storage was internal to the server, no
such Input/Output (I/O) interface was required, since storage was not externally
attached. With the advent of external storage, I/O interfaces (i.e., HBAs) became
necessary to provide inter-device communication.
Below in Figure 180, we have outlined the placement of NICs and HBAs within
server-to-PC and server-to-storage architectures. Note: since internal storage is
inside the server, there is no such need for an I/O interface.
Figure 180. HBAs Emerged as a Server-to-Storage I/O Interconnect
LAN Server
SAN Storage
Subsystem
PC
Fibre Channel
Ethernet
NIC HBA
179
The SAN Book III – October 22, 2001
Handled Handled
Function by HBA by NIC
Framing packets ✔ ✔
Provide MAC address ✔ ✔
Link level error checking ✔ ✔
Sequence level error checking ✔
Manages flow control ✔
Performs SAR ✔ Requires
Server
Provides linking of I/O request and packet ✔ CPU
Handles chaining of multiple requests ✔
Manages many concurrent I/O ✔
Ability to analyze and manage data ✔
Storage virtualization ✔
Source: Emulex
180
The SAN Book III – October 22, 2001
Below in Figure 182 we show the relationship between the host controller and the
target controller.
Figure 182. Host and Target Controllers Negotiate Data Transfer
Server Storage
Subsystem
Disk Drives
Fibre Channel
TARGET SCSI / FC
CONTROLLER
HOST SCSI / FC
HBA CONTROLLER
Source: Salomon Smith Barney
Below in Figure 183 are some examples of an HBA and host and target controller
chips from QLogic.
Figure 183. I/O Interfaces Exist in the Form of Chips, Boards, and Cells
181
The SAN Book III – October 22, 2001
LDO
FLASH
SRAM
OSCILLATOR
QLA2200 ASIC
182
The SAN Book III – October 22, 2001
Figure 185. HBAs Slide into the PCI Slot, Which Is Connected to the Server or PC’s Backplane (Left)
S-Bus Connector
in SPARC System
PCI Connectors
in PC
183
The SAN Book III – October 22, 2001
SBus Interconnect
PCI Interconnect
Source: JNI
184
The SAN Book III – October 22, 2001
Figure 187. QLogic and JNI Designs Have Similar Data Paths
GBIC GBIC
OPTIONAL OPTIONAL
TO FIBRE SRAM TO FIBRE SRAM
CHANNEL CHANNEL
CABLE SERDES CABLE
F
F I
I F
F SERDES O
O RISC RISC
PCI DMA
JNI is shifting from a While JNI has previously based its HBAs on Agilent’s Tachyon ASIC, in order to
licensed ASIC to its gain more control of its product roadmap and lower production costs it is
internally-developed
“Emerald” platform.
transitioning to its own ASIC family called Emerald. Interestingly, the Emerald
technology was originally sold to JNI by Adaptec, which dislodged its Fibre
Channel division in 1998. JNI’s new chip release, Emerald, serves as the basis for
nearly all of its new products, including PCI-X. By owning its own ASIC
architecture, JNI will be better served as the Fibre Channel market evolves to 2Gbps,
dual-channel, and support for Virtual Interface (VI), Internet Protocol (IP), and
iSCSI. In addition, a strong semiconductor design team will be crucial for JNI if it
decides to formulate a chip strategy for future growth.
Emulex Takes the High Road
The Emulex design is not necessarily centered around silicon integration, but around
performance (including reliability and functionality) and software instead. As
shown in Figure 188 below, Emulex uses an Intel StrongARM processor, which runs
up to 200–250 MHz, to manage the data flow. The benefit of Emulex’s design is
that it can better and more quickly leverage industry standard microprocessor
advances. Also, Emulex’s higher chip count circumvents distant limitations and
increases signal integrity (both of which are important in high-end and SAN
environments).
The downside to the chip-intensive solution is that a high part count increases costs
and takes up a large footprint; thereby requiring significant reengineering to migrate
to an embed solution for the motherboard (i.e., a chip solution). The higher part
count also gives Emulex a high-cost structure. As a result, Emulex’s gross margins
trend around 50%, less than QLogic and JNI which have historical gross margins
between 55% and 65%.
Another benefit of Emulex’s design is the software/firmware platform, which as a
result of the standardized hardware design, handles much of the data flow and
control and includes support for Fibre Channel and FICON on every board.
Integrating these functions in software, rather than silicon (as QLogic and JNI have),
provides greater flexibility for Emulex customers. As a result of Emulex’s chip-
intensive solution, it is well respected in the industry as the premium provider.
185
The SAN Book III – October 22, 2001
Figure 188. Emulex LP8000 and Block Diagram Showing Data Path
LDO FLASH
SRAM
Emulex Fibre Channel HBA
StrongARM
LDO
GBIC
TO FIBRE SRAM
CHANNEL
CABLE
DragonFly
SRAM SERDES
StrongARM
SerDes
Emulex’s flexible Emulex’s HBA architecture is ideal for high-end OEM customers, who often
architecture is ideal for implement proprietary features in their products. As a result, IBM uses Emulex’s
high-end customers.
products throughout several platforms including Sequent NUMA, RS6000, and
S390. In addition, IBM’s next generation FICON based mainframe system, referred
to as the zSeries (previously code-named Freeway), uses Emulex HBAs in both the
front-end server and back-end Shark storage subsystem. IBM has the only
subsystem that uses an HBA rather than a target chip. This means Emulex gets a
twofer in captive IBM server-to-storage environments by putting an HBA on both
ends of the connection.
In addition, Emulex’s design bug fixes are easily applied in its firmware which is
helpful when problems arise; and bugs are not uncommon. For example, in
September, 2000, Emulex customers discovered a bug in their LP8000 HBA, which
the company fixed and corrected in its software. As a result, shipment of new
product was only put on hold for a week and existing customers downloaded a
simple software update from Emulex’s website. This could have been a much more
costly problem if Emulex used a more integrated design.
High-end...Low-end?
Although we attribute Emulex’s design to the high-end market and QLogic’s to the
volume market, both have been making strides to cover the other’s so-called market.
In the end, it is more about winning customers. The first vendor to win a customer
can establish significant barriers to entry due to server and storage companies’ tight
development and integration to specific HBA designs (be it the high-end or volume
market). Note: It takes a new HBA design many months to get OEM approval, even
from an existing supplier using the same drivers.
Note: While QLogic supplies HBAs into Sun’s high-end products, Emulex supplies
to EMC and IBM’s volume NT and NAS products. Integration by vendor is often
more important than by platform.
The first mover advantage is almost everything in the HBA market. Emulex has
many first-mover design wins, including IBM, EMC, and Compaq. Qlogic has the
first mover advantage at Sun, Dell and NetApp. JNI for EMC (Sbus) and HDS.
Thus far, its has proved to be very difficult (and very unsuccessful) to try unhooking
186
The SAN Book III – October 22, 2001
an incumbent HBA supplier. Clearly this market cannot commoditize until HBAs
from different vendors are seamlessly interchangeable (or “plug and play”) which
we expect to take some time.
The Emerging HBA’int Market: a Single-Chip Solution
QLogic and JNI have the If Fibre Channel follows in the footsteps of the IDE and SCSI interfaces, a new
highest integration levels market segment will emerge: a chip based (silicon) solution for motherboards. We
.
believe this migration is crucial to promote broader Fibre Channel adoptions by
helping it penetrate the lower end of the market that is not as performance sensitive
and is more price sensitive. To date, for the 1 Gbps generation of Fibre Channel
products, QLogic is the supplier with the highest degree of integration with on-board
SerDes and RISC processor in its ASIC; the integration enables QLogic to have the
highest gross margins and the lowest ASPs. This low cost has enabled QLogic to
penetrate the entry-level Windows NT markets.
JNI’s Emerald-IV ASIC also integrates about every part needed to incorporate a
single chip solution, except the SerDes. JNI chose to avoid the complexities of
SerDes integration, recognizing that vendors such as Vitesse sell the 1 Gb part
(which is generally manufactured in GaAs) for only $5.
Trade-offs with an At the end of the day, while single chip solutions might be ideal for some
integrated SerDes. applications, they will not cover the mid- and high-end markets. For example, in our
opinion, signal clarity is highest with an external SerDes and larger SRAM buffers.
Therefore, full-blown SAN applications requiring external cabinet connections often
prefer the cost/performance trade-off of the higher-end HBAs (such as Emulex’s).
However, for smaller SANs and especially NAS applications, which may only
require inter-cabinet storage, we expect a more integrated HBA or chip solution
(such as QLogic’s) to be preferred.
Competitive Landscape
We expect market conditions to become more competitive; providing increased
share to the stronger players (the gorilla game). Already, strong players such as
Emulex, JNI, and QLogic have demonstrated solid growth rates, while smaller
players have underperformed. In addition, traditional SCSI providers, with the
exception of QLogic, have failed to gain traction in the Fibre Channel HBA market.
Below in Figure 189, we plot the various growth trajectories of the three leading
Fibre Channel HBA vendors: Emulex, JNI, and QLogic.
187
The SAN Book III – October 22, 2001
Figure 189. Historical Growth of Fibre Channel HBA Revenues per Quarter
Revenues in millions
$80
$70
$60
$50
$40
$30
$20
$10
$0
Mar-99 Jun-99 Sep-99 Dec-99 Mar-00 Jun-00 Sep-00 Dec-00 Mar-01 Jun-01
Players’ Positions
The various market positions of each of the Fibre Channel HBA vendors are
revealed by a study of their respective market shares. A recent IDC study of 2000
results (shown in Figure 190 below) proclaimed Emulex the unit and revenue share
leader achieving 32.1% and 34.8% of these respective markets. We attribute
Emulex’s strong showing to its focus on a high-end design solution which spans the
UNIX, NT, and mainframe platforms, translating into a broader footprint and higher
ASPs.
The study also shows that in 2000, QLogic was able to maintain a disproportionate
share of the units shipped when compared to its revenue share. QLogic’s unit and
revenue market shares were 29.1% and 17.0%, respectively. QLogic focuses more
closely on the price/performance market and does not command the same pricing
premium as Emulex. However, it does a similar unit volume and if QLogic’s host
and target chips were included we believe that QLogic would be the unit market
share leader.
JNI has an interesting position in the market: most of its shipments have been for
Solaris systems with about 70% of the company’s revenues from Solaris SBus
adapters. Also, the company’s close relationship with EMC has been lucrative given
the popularity of the Sun server/EMC storage configuration.
188
The SAN Book III – October 22, 2001
Agilent
12%
Qlogic
Qlogic JNI
17%
29% 18%
Agilent’s Tachyon ASIC Agilent was one of the early developers of Fibre Channel technology; however, by
is used in several early our analysis, the company’s success appears somewhat understated in the HBA
generation host
adapters.
market share numbers. The market share data show that Agilent has not had much
success in the HBA market, however, its ASIC (called “Tachyon”) development
efforts landed the company several design wins on several early generation HBAs.
The Tachyon hardware is based on a state machine architecture, which allows
performance of the HBA to scale proportionally with the system CPU.
Below in Figure 191, we illustrate some board and silicon relationships.
Figure 191. Board/Silicon Relationships
ADAPTEC COMPAQ EMULEX INTERPHASE JNI QLOGIC SUN MICRO
SUN
EMULEX
AGILENT AGILENT AGILENT ASIC
DRAGONFLY JNI QLOGIC
TACHYON TACHYON TACHYON or
or CENTAUR EMERALD ISP2x00
ASIC ASIC ASIC QLOGIC
ASIC
ASIC
189
The SAN Book III – October 22, 2001
suppliers, independent suppliers such as Emulex, JNI, and QLogic should be able to
easily outpace industry growth rates.
While the independent HBA suppliers (Emulex, JNI and QLogic) significantly
increased their combined market share from 39% in 1998 to 70% in 2000, the
captive suppliers (Compaq and Sun), which declined in market share to 18% in 2000
from 53% in 1998, have clearly begun to pull out.
Figure 192. Market Share of FC Host Bust Adapter Suppliers
1998 2000
Other Interphase
Interphase Agilent Sun Other
0% 3%
8% 0% 6% 2%
Qlogic
Agilent
JNI 9%
7% Emulex
9% Sun 35%
Compaq
43%
12%
Compaq
10%
Qlogic
Emulex JNI
17%
21% 18%
Name Association
Because the Fibre Channel HBA resides within a server, and ultimately connects to
a storage subsystem, it is a critical determinant of the overall system performance.
Therefore, end users generally select the HBA that has been approved by either the
server vendor, such as Compaq and Sun, or, the storage vendor, such as EMC,
Compaq StorageWorks, or Hitachi Data — or both! To make matters more
complex, the HBA can also be procured either from the server vendor, storage
vendor, or through indirect distribution channels. The following figure illustrates
some supplier/customer relationships for Emulex, QLogic, and JNI.
190
The SAN Book III – October 22, 2001
Emulex
OEM 85% & NT) , HDS (all),
Fujitsu (Solaris),
McData (all)
Bell, Info-X, Tech Data,
Channel 15% AVNET, Tidalwire, ACAL
Which relationships are most important? In our view, clearly, the largest server and
storage OEMs are most influential. One popular combination has been to combine
Sun Solaris servers, NT servers from Compaq or Dell, and EMC storage units, all
connected using a Brocade switch. Resellers would likely recommend a JNI HBA
be used for the Sun server, since EMC has tested and qualified JNI’s SBus HBA.
Sun and other vendors may offer an HBA, but unless it has been certified by EMC
also, the use of it may void the EMC warranties. Needless to say, Compaq, IBM,
Dell, MTI, and other major OEMs have similar programs.
191
The SAN Book III – October 22, 2001
storage sales, Emulex often gets a twofer, since it also embeds an HBA in IBM’s
Shark in those environments.
Below we have drawn a rough sketch of Emulex’s platform penetrations based on
our own in-depth industry research. Note: A good portion of Emulex’s revenues
come from the storage side within Shark which we have assigned to the AIX and NT
platforms. In the future, FICON should emerge as a noticeable contributor for
Emulex. We have based our estimates on HBA unit volumes.
Figure 194. Emulex’s Platform Profile
NAS
Tru64
1%
4%
AIX
31% NT
46%
Solaris
18%
Solaris
JNI has captured most of Sun Micro’s version of UNIX, called Solaris, is used mainly on Sun and Fujitsu
the HBA market for (which licenses a version of Solaris from Sun) servers. Emulex, JNI, QLogic, and
Solaris SBus, which is
commonly used in Sun’s
Sun each have Solaris drivers available for HBAs built for the PCI bus, which is
high-end servers. used in entry-level and workgroup servers from Sun.
Emulex, JNI, Sun and (recently) QLogic each have Solaris drivers available for
SBus HBAs, though JNI is recognized as the leading provider of SBus HBAs for
Solaris. QLogic has historically had the closest relationship with Sun including
some embedded chip design wins. We expect QLogic to continue to further
penetrate Sun.
Below in Figures 195 and 196, we have drawn a rough sketch of QLogic and JNI’s
respective platform penetrations based on our own industry research. We have
based our estimates on HBA unit volumes.
192
The SAN Book III – October 22, 2001
NAS
Solaris 9%
12%
NT
79%
Solaris
90%
Windows NT/2000
QLogic supplies major QLogic is the predominant supplier of Fibre Channel HBAs for the high-volume
NT server vendors such Windows NT server market. QLogic’s mini-port driver, which is commonly used
as Dell and IBM.
for NT systems, is highly regarded in the industry. As such, QLogic has captured
much market share in this segment and has relationships with OEMs such as Dell
and IBM’s Netfinity group. Adaptec provides HBAs to Hewlett-Packard for its
NTserver platforms, while Emulex is a supplier to Compaq for its Proliant NT
servers.
193
The SAN Book III – October 22, 2001
$14,000
$12,000 75.4% CAGR
$10,000
$8,000
$6,000
$4,000
$2,000
$0
2000 2001 2002 2003 2004 2005
194
The SAN Book III – October 22, 2001
Fibre Fibre
Server Channel Channel
Switch
Server
Source: Salomon Smith Barney
195
The SAN Book III – October 22, 2001
fabric communications are translated back to the legacy server for a seamless two-
way communication. Since only the setup is done in software (actually firmware),
once loaded, the translation occurs in the ASIC hardware and performance can be
sustained at full Fibre Channel speeds.
Gadzoox has approached this in a unique way by developing its Capellix chassis-
based switch. The chassis part means that it is modular in order to plug either a FC-
AL switching blade or fabric switching blade into a slot, thus making the switch
perform with the desired protocol. This solution requires that the system be taken
down (including unscrewing the lid of the switch) in order to upgrade. In our
opinion, this process is just about the last thing an IT manager wants to do.
Switch Addressing — Public Versus Private Loops
Hubs and FC-AL Hubs and FC-AL switches use simple local addressing schemes, much like an office
switches use simple phone system. To dial a person across the floor, you key in four digits. You don’t
local addressing
schemes. Fabric
have to dial the whole number because you’re in a local network. These devices are
switches and Directors referred to as “private loops,” because they can only see and communicate with
use more advanced other devices within their private loop.
(public) addressing
technology. Fabric switches use a more robust addressing scheme. These devices can be thought
of as your telephone at home. To transmit your voice, you dial the three-digit area
code, the three-digit zone, and then the four-digit address. This tells the phone
switch what area to go to, what zone to go to, and then what exact address to go to
within that zone. This is the same way a SAN fabric switch operates. Fabric
switches are referred to as “public” because they can see and communicate with
most every device connected to the network.
Switching Architectures
We believe that software Different vendors have architected their switching products with varying
remains one of the most components, cost structures, reliability, scalability, flexibility, and sophistication.
important features when
assessing the
While switching hardware architectures are wildly different (and we’ll go through
competitiveness of the range), we believe that software remains one of the most important features
assorted products. when assessing the competitiveness of assorted products. Software leverages
hardware to provide customers with solutions. We believe McData’s recent
acquisition of SAN Navigator (a storage software company focused on network
management) is an outcrop of the demand for software functionality. It has been our
observation that the best solution, not the best hardware, grabs the most market
share.
Switching Interconnects
Similar to our discussion of storage subsystem backplanes and PCI bus versus
Infiband, a switched interconnect can significantly increase performance. Using a
crossbar switch architecture, however, demands more logic, which adds complexity,
increases costs and potentially delays a product’s time to market.
Below in Figure 199 we illustrate a few different switch (or Director) architectures,
including Bus (Brocade and QLogic), switched (McData), multi-stage (Inrange), and
passive (Brocade’s Silkworm 12000).
196
The SAN Book III – October 22, 2001
ASIC ASIC
ASIC ASIC
ASIC ASIC
ASIC
ASIC ASIC
Brocade
Brocade uses a bus interconnect and leads the market in product maturity and ASIC
functionality. Brocade’s developments — trunking, time stamping, frame filtering
and traffic monitoring — are among the industry’s leading functions. Brocade’s
focus on real, end-user needs and a “total solution” design has enabled it to maintain
its popularity with end users.
197
The SAN Book III – October 22, 2001
RS-232
10/100 Ethernet
P ow
er
SFP
From Fibre
Channel
Buffer
Cable SerDes
(send)
Pentium
Class CPU
Controller DRAM
Flash
SerDes
SFP
To Fibre
Channel
Cable
(receive) SerDes
Memory
Buffer
McData
McData takes the highroad with a serial crossbar interconnect. The serial crossbar
enables any-to-any port connectivity without performance degradation. Since
McData uses a serial crossbar interconnect it must traverse two SerDes (or the same
one twice) per every send or receive transmission (or four per every send and receive
transmission). The extra SerDes workload is brought on due to the “serial”
technology within the crossbar which requires a serial transmission versus the ASIC
which requires a parallel transmission. For this reason, McData uses two SerDes per
ASIC (each SerDes used by McData has four quad channels). The SerDes (which
stands for Serializer/Deserializer) transforms the signal into a serial transmission
(Serializer) and/or into a parallel transmission (Deserializer).
McData uses a serial crossbar interconnect in order to increase performance and
utilize off-the-shelf technology (in this case, from Vitesse).
McData has developed a four-port ASIC design and has the highest component
count in the industry, which means it also has the highest cost structure. The benefit
is that McData also has the most reliable design. If one of its ASICs fails, it only
198
The SAN Book III – October 22, 2001
takes down four ports, versus higher ASIC densities, which can take down as many
as 16 ports upon failure.
Below in Figure 201, we have laid out a 16-port McData switch (ES-3016) and
mapped a data send and receive transmission in order to make sense of this
gibberish.
Figure 201. McData’s 2 Gbps 16-port Switch Architecture
RS-232
10/100 Ethernet
Buffer
Pow
SFP er
From Fibre
Channel Flash
SerDes
Cable SerDes
(send)
DRAM
SFP
To Fibre
Channel Serial
SerDes
Cable SerDes Crossbar
(receive) Controller
SerDes
SerDes
McData’s 4-
port Viper
ASIC
199
The SAN Book III – October 22, 2001
embedded crossbar technology inside its new Itasca Fibre Channel chip. As with its
HBAs, this provides QLogic with the most compact and lowest cost design in the
industry.
Pow
er RS-232
10/100 Ethernet
SFP
From Fibre SerDes
Channel
Cable
(send)
SerDes
GL_Port 0 Pentium
Class CPU
SerDes
Controller
DRAM
SerDes
Flash
GL_Port 11
Interconnect
Embedded
SerDes QLogic’s
16-Port
Itasca
SerDes
ASIC
SFP
To Fibre
Channel
Cable SerDes
(receive)
SerDes
200
The SAN Book III – October 22, 2001
port loss upon failures. While this might not be an issue for lower- and mid-range
environments, higher-end applications might prefer lower port densities per ASIC
(not to be confused with ports per switch or Director, they are very different
concepts that do not run parallel). Lower ASIC port densities are often viewed as
more reliable because if an ASIC does fail, it takes down a smaller fraction of the
overall switch or Director (i.e., if an ASIC fails, only four ports fail rather than eight
or 16). That said, ASIC failures are very unusual and redundancies can be built in
for greater reliability.
Other important aspects Other important aspects of ASICs are their functionality and software integration.
of ASICs are their While a customer might not pay as much attention to port count per ASIC, they
functionality and
software integration.
surely recognize the delivery of functionality. Functionality and software
integration are key elements to the successful delivery of a solution (i.e., what’s best
for the customer). For example, the ability of Brocade’s ASIC to autosense FC-AL
and fabric environments has enabled it to grab the lion’s share of the storage
networking market early on, as customers migrate from FC-AL to fabric. The
inability of other vendors to effectively mix these environments could have stalled
these vendors’ attempts to penetrate the market.
Other valuable, upcoming ASIC features include trunking, time stamping, frame
filtering, and traffic monitoring.
Trunking
Trunking, combined with load balancing, is one of the more anticipated inter-switch
functions. Trunking aggregates bandwidth to create a single, larger pipe that all data
can traverse. Load balancing equalizes the distribution of data across multiple
channels. Combining the two features results in better performance and less
congestion.
Below in Figure 203, we illustrate the way Trunking works when combined with
load balancing.
201
The SAN Book III – October 22, 2001
Without Trunking
2G 1G
2G
1.5G Congestion 1.5G
.5G .5G
1G 1G
2G 1G
2G With Trunking 2G
1.5G 1.5G
.5G .5G
1G 1G
8Gbps ISL Trunk:
2G All streams share bandwidth
2G
ASIC preserves
in-order frame delivery
Source: Brocade
Frame Filtering
Frame filtering is another ASIC capability that would integrate with available
software and enable features that would allow for successful system deployment.
With frame filtering, many features can be developed to improve management and
monitoring capabilities, and security in the SAN. Features such as LUN zoning,
detailed performance analysis, and multi-protocol zoning could improve the SAN
deployment and overall management, as well as ensure future integration with other
protocols.
202
The SAN Book III – October 22, 2001
$6,000
$4,000
$3,000
$2,000
$1,000
$0
2000 2001 2002 2003 2004 2005
Source: Dataquest
Hubs Directors
Directors 7% 10% Hubs
27%
31%
Switches Switches
66% 59%
Players
Brocade is the clear Brocade is the clear leader in the Fibre Channel switch segment, garnering an 86%
leader in the Fibre market share in 2000, according to IDC. Brocade has been able to maintain its
Channel switch segment.
switch market share at around 90% since the industry’s inception. We expect
McData, a long-time Director-only supplier, to play a greater role in the switch
203
The SAN Book III – October 22, 2001
market since the release of its 16- and 32-port switches, which began shipping at the
end of 2000, and were adopted by OEMs early in 2001. QLogic should also play a
bigger role in the future with its Sun relationship, early delivery of 2 Gbps and
integrated chip design, which should prove to be very leverageable in the embedded
switch market. Note: QLogic is the only vendor in both the Fibre Channel switch
and HBA markets.
Figure 206. Switch Market Share (Revenues) 2000
Vixel Gadzoox
QLogic 2%
3%
9%
Brocade
86%
$180
$160
$140
$120
$100
$80
$60
$40
$20
$0
Mar-99 Jun-99 Sep-99 Dec-99 Mar-00 Jun-00 Sep-00 Dec-00 Mar-01 Jun-01
204
The SAN Book III – October 22, 2001
Silkworm 12000
Silkworm 6400
(64-port core switch)
(Pre-configured
Silkworm 2400 64-port switch)
(8-port switch)
Silkworm 2800
(16-port switch)
Below in Figure 209, we also illustrate QLogic’s switch portfolio, including its
Director, which it OEMs from Inrange. QLogic has been successfully growing the
switch business it bought with its acquisition of Ancor in August, 2000. In
QLogic’s June 2001 quarter its switch business grew, to $14.4 million, or about 10%
sequentially, up from the $7–$9 million level when QLogic acquired it.
205
The SAN Book III – October 22, 2001
What Is a Director?
Think of Directors as Defining a Director is like defining the perfect cheeseburger: It depends on who you
high-end, fully redundant ask. By definition, Directors are high-end storage switches.
switches.
We have found that the most common criteria for Directors include:
➤ high port density
➤ scalability
➤ fully redundant, fabric, non-blocking architecture that is hot swapable
It is important to keep in mind that there are two very different types of storage
Directors: ESCON/FICON and Fibre Channel.
ESCON and FICON Directors (mainframe)
ESCON (mainframe) ESCON (mainframe) Directors have been in use for almost a decade. ESCON
Directors have been in Directors provide mainframe connectivity; in other words, in order to connect
use for almost a decade.
multiple storage subsystems to a mainframe one must use an ESCON Director.
Mainframes communicate through ESCON (Enterprise System CONnection), much
the same way that open system servers and storage subsystems communicate using
SCSI.
206
The SAN Book III – October 22, 2001
ESCON/FICON Directors have been built for zero downtime (or five nines of
reliability, meaning 99.999% uptime). In order to accomplish this goal, Directors
have been designed with multiple hardware redundancies, including power supplies,
fans, backplanes, processors, and memory, as well as the software functionality (i.e.,
logic) to utilize its fully redundant architecture in case of component failures.
Industry sources indicate that McData had an approximate 85% share of the ESCON
market in 1999 by supplying IBM’s 9032 Director that scales to 248 ports per unit.
Inrange captured about 15% of the market with its CD9000 that scales to 256 ports
per unit.
Note: IBM’s FICON products launched in the fall 2001.
Fibre Channel Directors (Open Systems: UNIX and NT)
Inrange and McData Since Inrange and McData (a former subsidiary of EMC) already had
entered the Fibre ESCON/FICON expertise, when they entered the open systems storage market, both
Channel market with
Director designs.
companies approached it from their Director heritage; thereby creating a new
category of Fibre Channel switches called Directors. (Directors are typically about
twice as expensive as Fibre Channel switches on a per-port basis; switch ASPs are
about $1300 per port; Director ASPs are about $2500–$3000 per port). McData’s
first generation 32-port Fibre Channel Directors (the ED-5000) began shipping in
quantity in October, 1998, and Inrange introduced its 64-port Director (the FC/9000)
in April, 2000.
Director...Switch...So what the Heck Is the Difference?
Thus far, the market has essentially defined Directors as high port density switches.
McData’s Directors had 32 ports; Inrange’s 64 ports. Directors, however, are much
more than simply high port densities: They are fully hardware redundant Fibre
Channel fabric switching devices that can maintain full non-blocking bandwidth in
the event of multiple hardware failures. In contrast, switches would have to
substantially over-allocate ports in order to ensure a non-blocking, fully available
and reliable architecture. But, the resulting switch mesh would result in a higher
cost and a more complex, harder to manage environment.
Baby You Can Drive My Car
We liken the technology to a car with four tires going down the highway at 60 mph.
Then one tire blows out (i.e., hardware failure). The Director drops down a
replacement tire and keeps on trucking at 60 mph uninterrupted (like a James Bond
207
The SAN Book III – October 22, 2001
car). The software-redundant switch, on the other hand, redistributes (balances) the
weight of the car such that it can run on the remaining three tires until the fourth tire
can be replaced. This type of switching architecture might suffer some performance
degradation and have to reduce its speed in order to maintain total throughput.
Now imagine that the car has 64 or 128 tires. One tire blowing out is a lot less
significant. In other words, the larger the fabric, the less important one port (tire)
becomes. However, while we believe a software redundant “core switch” might be
able to service a good portion of the market, in our opinion, many IT managers will
simply demand hardware redundancies in order to be conservative. That said, we
believe that switches (including 16-port switches) will continue to dominate the
market, from the point of view of total ports shipped, for at least another couple of
years. As Brocade often points out and we agree, the issue will continue to be about
the size of your overall fabric and how its managed.
Source: Dataquest
Players
In 2000, McData was the clear Director leader with an 85.4% market share.
However, Inrange definitely deserves the “up-and-comer” award for establishing a
14.6% share after only having product available in the second half of the year.
208
The SAN Book III – October 22, 2001
Inrange
15%
McData
85%
Inrange
Inrange entered the Fibre Channel Director market in April, 2000, with its 64-port
FC/9000 (the largest density at the time) and has since scaled its Director up to 128
ports (the largest density in the market). Inrange uses QLogic’s (formerly Ancor’s)
Fibre Channel ASIC technology in its Director and this has created a tight
partnership between the two companies. Inrange OEMs QLogic’s eight- and 16-port
switches and QLogic OEMs Inrange’s FC/9000 Director.
Inrange has also delivered its own ESCON and FICON Directors, which it sells
through its direct sales force, resellers, and OEMs. The company has also stated its
intention to migrate to a 256-port Director by 2002 (current 64- and 128-port
FC/9000s are architected to be fully upgradable to 256 full bandwidth, non-blocking
ports). Inrange can offer Fibre Channel (fabric and FC-AL, both public and private)
and FICON connectivity within a single box. The FC/9000 also supports IP-over-
Fibre Channel.
McData
McData entered the Fibre Channel market in the fourth quarter of 1998 with its 32-
port ED-5000 Director, which used third-party ASIC technology. In November,
2000, McData delivered its own ASIC technology and introduced its 16-port ES-
3016 switch, its 32-port ES-3032 (March, 2001) and its 64-port ED-6064 (April,
2001). McData also manufactures and designs its own line of ESCON Directors,
ESCON-to-FICON bridge cards (which fit into its ESCON Director), and FICON
Directors.
McData’s new generation 16-, 32- and 64-port products all support Fibre Channel
fabric and FICON in the same enclosure and provide the industry’s most densely
packaged solution (i.e., offering the most ports per square inch). Note: McData does
not support FC-AL in its switches or Directors, except in its ES-1000, which works
in FC-AL mode but not in fabric mode.
Brocade
Brocade is planning to enter the Director market with its Silkworm 12000 late this
year. The Silkworm 12000 is a 64- and 128-port “core switch” that provides much
209
The SAN Book III – October 22, 2001
of the same reliability as a Director though it does not, however, support FICON
(i.e., the mainframe). We expect the Silkworm 12000 to drive significant growth for
Brocade in 2002 and beyond as it opens up the Director market and enables the
company to expand its footprint.
The figure below gives a rough timeline of the release of many Fibre Channel
switching technologies.
Figure 213. Fibre Channel Product Introductions
Brocade &
2Gbps
QLogic Inrange McData
Brocade to
McData 16- introduce 64-port
Brocade 8/16-port port ES-3016 Silkworm 12000
Silkworm
2400/2800
Brocade 64-port
Silkworm 6400
210
The SAN Book III – October 22, 2001
FC/9000
(64 ports)
FC/9000
(16 ports)
Source: Inrange
211
The SAN Book III – October 22, 2001
nodes, but only uses three 16-port switches (48 ports). However, the latter example
is not a fully redundant, non-blocking architecture, meaning that performance can
degrade as port failures occur.
Note: The diagrams below are also illustrations of a multistage or meshed
architecture.
Figure 215. Building a 32-port High Availability Fabric Using 16-port Switches
16-port Switches
8 nodes
8 nodes
8 nodes
8 nodes
*Note: there are less expensive/complex ways to build a 32-port fabric. However, the less
expensive solution will not be as robust. We architected the above design to compare apples to
apples on the high-end without sacrificing performance and availability.
16-port Switches
10 nodes
10 nodes
10 nodes
212
The SAN Book III – October 22, 2001
Storage
Servers
Switches Switches
Directors or
Core Switches
Core
Edge Edge
Thus far, McData has been the leader in the core-to-edge market due to its time to
market delivery of a full solution spanning Directors and switches. However, we
expect Brocade and Inrange to establish solid footholds in this segment.
Figure 218 shows a more likely core-to-edge architecture.
213
The SAN Book III – October 22, 2001
Edge
Core
ES-1000 ES-3016
ES-3032
ED-6064
Source: McData
Compatibility Matters!
Interoperability is a key Although many products seem robust, when you get into actual computing
component to promoting environments involving multiple vendors and platforms, they fall a bit short due to
broad SAN adoption.
incompatibility. In other words, what good is a switch that cannot connect to
214
The SAN Book III – October 22, 2001
215
The SAN Book III – October 22, 2001
— are an integral part of bringing the concept of Fibre Channel SANs to reality
today.
Bridges can be in stand-alone boxes or integrated within storage subsystems (and in
which case are called blades). Figure 220 shows a SAN bridge.
Figure 220. SAN Bridge Diagram
Bridge
Source: Salomon Smith Barney
216
The SAN Book III – October 22, 2001
cemented. Note: In Dataquest’s analysis and projections, bridges and routers are
placed in the same category.
Figure 221. SAN Router Diagram
Players
Crossroads has been the Crossroads has been the clear leader in the SAN router segment, although there are
clear leader in the SAN other emerging niche companies entering the market, such as Chaparral, ATTO and
router segment.
Pathlight (recently bought by ADIC).
Figure 222. Crossroads’ Product Portfolio
Crossroads 8000
Front View
Rear View
Source: Crossroads
217
The SAN Book III – October 22, 2001
companies such as Brocade, Inrange, and McData have partnered up with leading
DWDM companies OpticalNetworks and ADVA, respectively).
Figure 223. Routers Could Enable Multiple Protocol Connectivity
Fibre
Ethernet Channel
Disk
Array SAN Routers can
enable the
interconnectivity
of multiple
protocols.
Router
Tape
Library
Source: Salomon Smith Barney
We believe storage routing could become a very large portion of overall storage
networking revenues (although it may become tough to distinguish this segment as
storage per se). We believe this market will gain traction in 2002 to provide SAN to
LAN, WAN, and MAN connectivity.
218
The SAN Book III – October 22, 2001
FC-GE
Router
Server 1 Disk 1
SAN
Disk 2
Server 2
Tape 1
LAN Server 3
Disk 3
Server 4 Disk 4
Disk 5
Server 5
SAN
Disk 6
Server 6
NAS Disk 7
Server 1
NAS
FC-GE
Server 2 Disk 8
Router
Players
Cisco entered into a technology license agreement with Brocade in June, 2000.
Under the terms of the agreement Cisco’s Catalyst 6000 product family of
multilayer switching technology (Ethernet) would be combined with Brocade’s
Silkworm switch technology (Fibre Channel). The interconnection would occur
through the integration of a Fibre Channel–based interface, designed by Brocade and
built by Cisco, located within a Cisco switch, providing switch-to-switch
connectivity. The combined solution is based on encapsulating Fibre Channel’s
protocol, FCP, over TCP stacked with IP through Ethernet to achieve non-blocking
OC48 performance (about 248 MBps). Cisco plans to continue increasing the speed
of transmission with future generations of the interface.
219
The SAN Book III – October 22, 2001
In our opinion, other leading companies in the SAN router segment include ADVA
(through its purchase of SAN Ltd.), Crossroads, Computer Network Technologies,
Nishan, and SAN Valley.
Figure 225. Nishan’s FC-to-IP Router
Source: Nishan
220
The SAN Book III – October 22, 2001
Fibre Fibre
Server Channel Channel
Hub
Server
Source: Salomon Smith Barney
221
The SAN Book III – October 22, 2001
222
The SAN Book III – October 22, 2001
Hub Switch
Figure 230. Six-port Hub and Switch Architectures (16-port Switches Are More Common)
Hub Switch
Note: The two previous figures were drawn for simplicity; in reality, any port in a
switch can connect to any other port. In our illustrations, switch ports only connect
across.
Notice that all the hub ports are connected to each other (called daisy chaining). In
order for one port to connect to another port in a hub, it must pass through all other
ports along the way. In a switch, various paths are available to simultaneously
connect multiple ports without interfering with one another creating point-to-point
connectivity.
223
The SAN Book III – October 22, 2001
Looplet 1
Switch
Looplet 2
Hub or
Switch
224
The SAN Book III – October 22, 2001
Protocols
225
The SAN Book III – October 22, 2001
z
Protocols: The Language of SANs
➤ While Fibre Channel continues to proliferate the industry in the form
of networked and non-networked storage, IP storage is getting closer
to becoming a reality. Several IP based products have been
announced recently, and the IETF is getting closer to selecting a final
standard. We expect IP storage to effectively penetrate the market
over the next three years, but instead of replacing Fibre Channel, we
believe it will complement it on the low end and at the periphery in
the near term.
WAN
LAN
SAN
226
The SAN Book III – October 22, 2001
On top of the seeming On top of the seeming sibling rivalries, newer protocols continue to show up.
sibling rivalries, newer InfiniBand appears one of the most promising. Over the next three years we are
protocols continue to
show up.
looking forward to watching InfiniBand’s evolution as a PCI bus replacement (i.e.,
an alternative I/O) and as a potential storage networking contender. However, the
most proliferated and aggressive technology has been IP (i.e., Ethernet; IP stands for
Internet Protocol), which has already crept into the WAN space by providing voice
over IP. IP is now positioning itself, through the recent introduction of iSCSI and
other iterations, to also enter storage networking.
Before diving into a protocol discussion, let’s first appreciate why so many different
vendors, technologies and investors are so intensely focused on this debate: Storage
networking is going to be a HUGE market and everyone wants a piece of it.
In this section we explore the various technologies and their respective potentials in
the storage market. In the process we will look to provide detailed explanations of
each technology and how we expect them to play out.
➤ Basics of I/O Interfaces – starts out in the weeds of system and connectivity
performance (I/O stands for Input/Output).
Various protocols can be used to network storage. We discuss 1) Fibre Channel, 2)
Ethernet, and 3) IP storage:
➤ Fibre Channel: Networking Storage — outlines the importance of Fibre
Channel and its advantages in storage networking.
➤ Fibre Channel versus Ethernet — Compares Fibre Channel and Ethernet and
tracks the histories of both.
➤ IP Storage — Comparisons of iSCSI, iFCP and FCIP, three IP-based protocols
gaining traction for storage networking with backing from influential companies
such as Cisco and IBM.
➤ InfiniBand — a new architecture that multiplies the maximum bandwidth of
interconnectivity in servers.
227
The SAN Book III – October 22, 2001
When?
In the end, we believe it We have heard many people implying that this concept of “one network” is just
will take at least two or around the corner and beginning to emerge today. While we agree it is being
three years before
Ethernet is ready to go
architected today, we believe the availability of a technology that can meet all types
head-to-head with Fibre of networking demand is in the distant future (more than five or ten years away).
Channel. Thereafter, we
believe the two It has been our experience that technology shifts of this magnitude take more time
technologies will — not less — than originally anticipated. Right now, people are talking about
overlap. Gigabit Ethernet moving to 10 GbE and 40 GbE; we note that Ethernet is currently
ramping transmission speeds of 1 Gbps. At the same time, it must be noted that
Fibre Channel is not standing still. 2 Gbps Fibre Channel has already hit General
Availability (GA) and the 10 Gbps Fibre Channel spec is already in the works.
In Figure 233 below we have depicted our expectations of the rollout of various
storage protocols.
Figure 233. Past and Potential Future Storage Networking Technology Availability
Adoption
el
nn
e
ag
ha
nd
tor
C
iba
S
re
fin
IP
Fib
In
1998 1999 2000 2001 2002 2003 2004 2005 2006
How?
The how is a bit tougher to explain. In the following sections we will explain the
difference between various networking protocols and what we believe will unfold in
the next three years. While a longer-term vision is exciting to talk about, we believe
it is too futuristic and difficult to predict and, therefore, should not consume
investors’ diligence today. Besides, the technologies will probably change at least
15 times in the next couple of years alone.
Learning from History
Who would have guessed that Cisco’s technology — a company that very few
people had heard of in the late eighties and early nineties — could have usurped Big
Blue’s efforts to network the Internet with its technology? The point is, nobody
really knows what technologies and companies will survive in the distant future (we
believe a more appropriate window to focus on is two or three years out).
228
The SAN Book III – October 22, 2001
Who Wins?
We certainly see many new technological advancements developing from all sides
We believe when the of the debate. For example, we believe Cisco’s torrid pace of acquisitions, including
dust settles in ten years, NuSpeed, is a clear indication that it recognizes the constantly changing network
we will have networking topology. We also believe almost every one of the Fibre Channel companies has
companies with Ethernet
roots and ones with
morphed quite a bit since inception. For example, Ancor, which was focused on
Fibre Channel roots. In Fibre Channel LANs in the early nineties, moved into Fibre Channel storage
our view, the key for networking, was bought buy QLogic, and is now developing iSCSI and InfiniBand.
investors is not to pick a Another good example is Inrange which has morphed from a telecom and datacom
technology, rather pick
quality companies which
networking company with some mainframe products to be a leader in the high-end
have displayed the Fibre Channel market. Inrange has long-haul IP and ATM connectivity and is
ability and willingness to looking to incorporate InfiniBand and iSCSI in future architectures.
adapt.
We believe the ultimate winners will be companies from both Ethernet and Fibre
Channel networking ancestries that have displayed 1) technology leadership, 2)
visionary management leadership, 3) the ability to react quickly to changing
technologies and changing customer demands, and 4) a willingness to evolve their
core competencies to suit customer needs two to three years out by developing fresh
technologies in new directions today.
Our Thoughts for the Future
➤ We believe the emerging “Data-centric” architecture (architecture, not network)
will incorporate multiple networking platforms, including Fibre Channel,
Ethernet, ATM, DWDM, and InfiniBand.
➤ Traditional networking companies will both compete and partner with Fibre
Channel networking companies well into the future.
➤ Ethernet will develop into a more robust technology over the next three years,
and only then, will it be able to begin competing head-to-head with Fibre
Channel to network storage.
➤ 10 Gbps Ethernet (10 GbE) will tip the scale toward using Ethernet, perhaps in
the form of iSCSI, as the preferred solution in the low end of the market and on
the periphery of storage networks.
➤ By the time Ethernet has developed the same level of functionality as Fibre
Channel, it will no longer be TCP/IP Ethernet. We believe it will require a
forklift upgrade (i.e., the installation of an entirely new IP network) with
hardware that can communicate with the new protocol, which we expect to be
different from today’s Ethernet hardware. We believe that iSCSI has taken the
lead as the most viable IP storage networking protocol, but expect that others
will continue to evolve as challengers.
➤ Fibre Channel will continue to be the unchallenged, premier storage networking
solution for enterprise class, production data for at least another two or three
years.
➤ Fibre Channel will continue to gain momentum through 2004, no matter how
many resources are stacked against it. Note: It takes a long time between a
229
The SAN Book III – October 22, 2001
technology’s inception and when it’s ready for prime time. It took Fibre
Channel, for example, more than five years to get ready for prime time.
➤ Storage Networking companies should continue to recognize record growth and
storage networking will likely expand into a HUGE market opportunity.
➤ InfiniBand should replace the PCI bus over the next two to three years and only
then will it begin to show up as a potential alternative to network storage.
➤ The Virtual Interface (VI) protocol will play an increasing role as systems get
more advanced and require faster I/O.
➤ Storage Networking companies have already begun and will continue to
incorporate Ethernet, ATM, InfiniBand, and any other networking technology
which presents itself as a viable storage networking alternative for the future.
230
The SAN Book III – October 22, 2001
What Is I/O?
Input/Output (I/O) I/O stands for Input/Output and describes the process and components involved in
describes the process sending and receiving data between computers and peripherals. For example,
and components
involved in sending and
computers can use the SCSI interface for its I/O with disk drives. SCSI is a protocol
receiving data between describing one way of sending and receiving data. Alternatively, computers can also
computers and use Fibre Channel or IDE.
peripherals.
Each protocol has its advantages and disadvantages. Hence, different ones are used
for different purposes. For example, while SCSI and Fibre Channel are used for
communication with disk drives, Universal Serial Bus (USB) protocol is often used
with communication for keyboards.
231
The SAN Book III – October 22, 2001
Cirque de Standards
Because of Intel’s significant presence in microprocessors, chipsets, and
motherboards, it has emerged as a pseudo-governor of computer architecture,
endorsing specific memory or graphics standards to ensure the PC “evolves” in an
orderly manner. To a degree, I/O standards have also been influenced by Intel. Its
endorsement of USB has shifted the low-performance peripherals market away from
serial and parallel ports. High-performance I/O, however, has been out of the
company’s jurisdiction, since Intel only determines specifications up to the PCI bus,
the portal through which many peripherals are attached. As to what devices are
attached to the PCI bus, it is the IT professional or systems engineer that determines
whether the I/O protocol will be IDE, SCSI, or Fibre Channel.
232
The SAN Book III – October 22, 2001
$300. In addition, disk drive manufacturers also charge a premium for SCSI drives.
To justify the cost premium, SCSI vendors have touted both speed and functionality.
The latest generation, Ultra320, has not only the highest data transfer rate of any
SCSI generation, but also new fault tolerance features such as Cyclic Redundancy
Checking (CRC) and Domain Validation. Both improve the integrity and reliability
of the data transfers. With Ultra320, SCSI has managed to keep pace with the
performance of current high-end computer systems (see Figure 234).
Figure 234. Evolution of SCSI
640
80 Ultra2 SCSI
MB/sec
LVD
40 Ultra Wide SCSI
20 Ultra SCSI
10 Fast SCSI
SCSI
5
1984 ‘86 ‘88 ‘90 ‘92 ‘94 ‘96 ‘98 2000 2001
233
The SAN Book III – October 22, 2001
remains a reliable and widely available interconnect technology, which has evolved
both in speed and connectivity over several years. Even though Fibre Channel is
gaining traction in the high-end applications, and IDE in low-end, the transition
won’t happen overnight. SCSI’s long history is difficult to overlook due to the fact
that engineers often go with what they have experience with.
For example, EMC has stayed with SCSI drives in its Symmetrix subsystem.
Adaptec, QLogic, and LSI Logic provide SCSI chips and HBAs.
Figure 235. Data Transfer Rates of Various Interface Standards
MBps
350
300
250
200
150
100
50
0 SI
SI
0
0
94
1G
2G
0
1.
2.
10
SC
SC
13
SB
SB
el
el
A
60
20
nn
nn
T
U
aA
a1
a3
ha
ha
eC
eC
ltr
ltr
ltr
U
U
br
br
Fi
Fi
234
The SAN Book III – October 22, 2001
1.0 has already become the next generation standard for low bandwidth I/O devices
(i.e., keyboards, mice). Already incorporated in current Intel chipsets, USB 1.0 has
lower data transfer rates than either IDE or SCSI and runs at 1.5 MBps. USB 2.0
will have significantly higher performance, up to 60 MBps. While USB 2.0 may
encroach on low-end storage solutions, such as desktop removable media (Iomega
Zip drives), it will not have the bandwidth to compete with the mainstream desktop
I/O standards.
InfiniBand
While this would be a great place to move to a discussion on InfiniBand, we feel it is
more appropriate (chronologically) to spend some time on Fibre Channel next, then
IP storage, and then InfiniBand.
235
The SAN Book III – October 22, 2001
Data 1
Control 1
Data 2 Multiple
Control 2 Wires
Data 3
Control 3
236
The SAN Book III – October 22, 2001
As one can imagine, it is difficult having 16 different pieces of data arrive at the
exact same time. SCSI has distance limitations: The longer the data transfer, the
higher the probability it will arrive at different times. Additionally, having many
electrical wires tightly wrapped together creates electromagnetic interference, which
can degrade the integrity of the data. Fibre Channel overcomes this distance
limitation by using a serial, instead of a parallel, interface. As a result, information
(still using the SCSI protocol) can be sent one bit at a time so the bits of data would
not have a problem in arriving in the exact same order in which they were sent,
regardless of distance.
Additionally, the architects of Fibre Channel added new features into the protocol
which extended its ability to be networked.
➤ High Bandwidth (Speed). The first generation of Fibre Channel debuted with
data transfer rates of 1 Gbps (or 1 Gbit/sec or 100 MB/sec or 100 MBps; it’s all
basically the same thing), faster than any other interface technology at the time.
Currently Fibre Channel’s second generation is being deployed and reaches
twice the data transfer rate, or 2 Gbps (or 200 MBps). This compares to SCSI,
which is currently running at 320 MBps, an increase from previous SCSI speeds
of 160 MBps and 80 MBps. Although Ultra 160 SCSI and Ultra 320 SCSI have
high bandwidth, as parallel protocols they lack several important features that
are useful for networking.
Another benefit of Fibre Channel is that it is bi-directional, which means that it
can send and receive data at the same time. This feature effectively doubles the
237
The SAN Book III – October 22, 2001
238
The SAN Book III – October 22, 2001
➤ Fibre Channel has won out over SCSI for large-scale SANs. The next
debate is whether Ethernet can beat Fibre Channel for future growth.
239
The SAN Book III – October 22, 2001
look at what today’s TCP/IP Ethernet is and compare it to today’s Fibre Channel.
Then let’s see whether, and how, Ethernet might work out in the house of storage.
What Is an Architecture?
Architecture refers to the Architecture refers to the hardware and software used to implement the protocol.
hardware and software For example, a network interface card (NIC) is part of the Ethernet architecture.
used to implement the
protocol.
Using Layers to Describe Protocols
What Are Layers?
Layer diagrams are often used to describe protocols. In order to send data across a
network, different layers of instructions must be added. As the data packet goes
through the different layers, it grows larger as more information is added.
Figure 238 below illustrates how data is networked using the OSI model which we
describe in further detail in the next section. Each discrete set of instructions is
represented by a separate layer in sequence. Note: Although OSI divides the
instructions into seven categories, any number of categories can be used.
240
The SAN Book III – October 22, 2001
241
The SAN Book III – October 22, 2001
OSI Fibre
Model Channel Ethernet
L7: Application
Data L6: Presentation
L5: Session
FC-4
L4: Transport TCP
L3: Network IP
FC-2
FC-1
Figure 240 below describes the type of processing that occurs on each layer.
Note: We have seen multiple ways of mapping Fibre Channel, Ethernet, and other
storage protocols to the OSI model from various respected sources. Since the OSI is
an idealized model which no one is required to adhere to, protocols do not always
match it exactly. For example, protocols can add new rules which are not included
in the OSI model, or they can group the layers differently. Since protocol mapping
is more of an art than a science, this represents our best approximation.
Figure 240. ISO’s OSI Model
Layer What it Does Analogy
L7) Application Translates the message into a form that Draft a letter.
the receiving computer will understand.
For example, text into ASCII.
L6) Presentation Translates, compresses and encrypts the Print out actual message on
message. stationery paper.
L5) Session Details about the communications Will it be sent first or second class?
session.
L4) Transport Segments the data. Breaking up the parcel into
separate packages.
L3) Network Adds addressing and routing information. Write down address and best way
to get there.
L2) Datalink Prepares data for particular method of If sending first class, put inside
physical medium. First class envelope.
L1) Physical Transfers from computer onto medium Send it out of the building via truck.
(i.e., copper, optical).
Source: Salomon Smith Barney
242
The SAN Book III – October 22, 2001
Sending data from host Sending data from host to target can be analogous to one person sending a letter to
to target can be another. Data begins to be processed from top to bottom, from layer seven to layer
analogous to one person
sending a letter to
one. The first three layers deal with the presentation of the data before it is broken
another. up into smaller packets of data. In layer seven, the computer transforms the raw data
into a form that can be sent, similar to writing your thoughts down in a draft letter
that can be read by someone else. In layer six, it formats it even further, similar to
transforming your draft into a nicely written letter that can be properly sent. In layer
five, it describes features of the communication session (i.e., security), similar to
deciding if your letter should be sent first or second class.
The next four layers deal with breaking up the data into smaller packets and
preparing these packets to be reassembled correctly upon arrival. In layer four, the
data is segmented into smaller pieces, similar to breaking down a parcel into
separate boxes. In layer three, the computer adds on addressing and routing
information, similar to writing down the address of the person you are sending your
letter to on the envelope, as well as the return address so the recipient knows who it
is coming from. In layer two, it prepares the data to be used properly depending on
what kind of physical medium it is travelling on, similar to putting the letter inside a
first class envelope so it can be sent correctly. In layer one, the data finally gets sent
onto the wire, similar to the post office putting the letter into the truck and out on the
road.
OSI Fibre
Model Channel
L7: Application
Data L6: Presentation
L5: Session
FC-4
L4: Transport
L3: Network
FC-2
FC-1
243
The SAN Book III – October 22, 2001
244
The SAN Book III – October 22, 2001
In the mid-1990s, when SCSI began to look as though it was running out of gas,
companies such as EMC and Brocade began to realize the benefits that Fibre
Channel could bring by networking storage.
OSI
Model Ethernet
L7: Application
Data L6: Presentation
L5: Session
L3: Network IP
L1: Physical
Figure 244 below briefly outlines the main function of each of these protocols in
Ethernet.
Figure 244. Ethernet — Protocol Functions
Layer Description
TCP Ensures reliable transport
IP Addressing
Ethernet Formatting data for physical transport
protocol
Source: Salomon Smith Barney
TCP/IP was originally design by the U.S. Department of Defense and largely used in
governmental agencies and universities. It was the use of TCP/IP to develop the
Internet, however, that gave it the ultimate upper hand over competitors such as
Novell’s SPX/IPX and IBM’s SNA protocols.
Ethernet Features
Features of Ethernet include: 1) different framing architecture, 2) longer instruction
sets, 3) transmitting individual packets on different routes versus sequences of
packets on the same route, and 4) different delivery guarantees.
245
The SAN Book III – October 22, 2001
If the data to be If the data to be transmitted is smaller in size, the packet is extended to 512 bytes
transmitted is small in through a technique called “carrier extension,” which adds extension symbols for
size, the packet is
extended to 512 bytes
collision detection purposes. The extra bytes are overhead and reduce overall
through a technique efficiency. Note: Ethernet jumbo frames allow for packets as large as 9000 bytes
called “carrier but are not commonly used or supported by all existing infrastructure.
extension.”
In Figure 246 below, the light grey areas have all of the relevant information that is
needed to send the data payload. However, the entire length of this data string is
less that 512 bytes. Hence extension symbols are added at the end to make sure the
packet is exactly 512 bytes long.
Figure 246. Ethernet Packet Adds Extension Symbols for Small Data Payloads
64 bytes
512 bytes
Fibre Channel is Fibre Channel, on the other hand, has frame sizes that range between 36 and 2,148
effective for high- bytes (that’s approximately 2 kilobytes). The amount of overhead, however,
volume, small frame
transmission, and also
remains constant regardless of the size of the data to be transmitted. This makes
has sequencing Fibre Channel highly efficient for sending high volumes of smaller frames.
capabilities for transfers
of large blocks of data. In addition, Fibre Channel frames can be linked together in a sequence, which makes
the protocol useful for large bulk transfers of data. Note: Gigabit Ethernet also has
frame bursting, which allows for the transmission of a series of frames at one time.
Figure 247 below illustrates a Fibre Channel frame. Unlike an Ethernet packet, the
Fibre Channel frame does not waste transmission cycles with extension symbols.
Both Ethernet and Fibre Channel frames contain the same type of information.
246
The SAN Book III – October 22, 2001
Note: Although the sections of the Fibre Channel frame in Figure 247 below are
labeled differently from those in the Ethernet packet above, both have the same type
of information, not including the extension symbols. The header includes the source
address, destination address and type; similar to Ethernet. The figure is not drawn
to scale as the data payload can be as large as 2112 bytes.
Figure 247. Fibre Channel Frames
Cyclic
Start of End of
Header Data Redundancy
Frame Frame
Check
36 to 2,148 bytes
2. Longer Instructions
Ethernet works well for small bits of information, but generates too much overhead
to effectively transmit larger blocks of data. It has a more cumbersome instruction
set (over 5000 instructions per stack versus a few hundred in Fibre Channel) which
requires the utilization of more CPU cycles than Fibre Channel to process a packet.
Additionally TCP/IP is typically processed using software. Fibre Channel’s
instructions are processed more quickly by using hardware (the HBA). Several
companies, however, have announced plans to develop TCP accelerators, which
process sections of the TCP/IP stack on hardware.
Also note that Gigabit Ethernet, although it raised its data transfer rate 10x by
moving to 1 Gbps, it only increased its throughput by approximately 3x as a result of
its high overhead (one could argue the increase is 2x or 4x, depending on the method
of measurement).
While TCP/IP does the 3. Sends Packets over Different Routes
job for file level data on
TCP/IP often routes each packet differently, giving each individual packet of a given
the LAN, it does not
currently meet the data transmission a different route to arrive at the same destination as the others;
performance whereas Fibre Channel logically groups frames and sends several at a time in
requirements of block sequence, ensuring they arrive in order. The result is that TCP/IP often drops
level storage
transmissions.
packets of data when the network becomes congested. The packets must then be
retransmitted using more bandwidth.
5. Guaranteed Delivery
TCP/IP always offers guaranteed delivery, a feature in which the receiver will send
an acknowledgement to the sender after each packet. If the sender does not get the
acknowledgement within a specified period of time, it will retransmit the packet.
247
The SAN Book III – October 22, 2001
Hence, when two stations are transmitting data on the same line at the same time, no
acknowledgement will be received by either. Each station will then wait a random
amount of time before transmitting again. In heavy traffic conditions, after ten
unsuccessful attempts, the delay time increases. After 16 unsuccessful attempts, the
station gets an error and no more tries are made. Although this reduces congestion
on a busy network, it will also slow down data transmission.
Fibre Channel offers six Fibre Channel offers six different classes of service that can change the delivery
different classes of service guarantee characteristics depending on the type of environment the SAN is in. Class
that can change the
delivery guarantee
1, a circuit-switched connection, is a dedicated, uninterruptible link, analogous to a
characteristics depending telephone connection. The benefit of Class 1 service is that transmissions are
on the type of environment sustained and guaranteed, which is useful for high-volume, time-critical
the SAN is in.
transmissions, such as between two large supercomputers. Class 2 service
guarantees delivery with receipt confirmation, but no dedicated connection is
established. Class 3 is similar to Class 2 but without a guarantee of delivery so there
is no wait for confirmation. Note: In Class 3, the most common implementation of
Fibre Channel, there is no guaranteed delivery, which means Fibre Channel
immediately resends packets upon receiving an error.
248
The SAN Book III – October 22, 2001
Introduction
The Internet Engineering Task Force (IETF), a standards body for Internet
architectures, is currently considering several standards for IP storage. We use the
term IP storage to describe protocols that are based on IP over Ethernet networks
versus Fibre Channel. Although network-attached storage (NAS) appliances also
work over Ethernet, we do not include NAS per se in our definition of IP storage.
Three of the more Three of the more prevalent IP storage protocols are iSCSI, iFCP, and FCIP.
prevalent IP storage Understanding the differences between these proposals can be helpful in comparing
protocols are iSCSI,
iFCP, and FCIP.
the unique benefits they have to offer.
In this section, we explain IP storage and what stages we believe this technology
will go through in the future.
249
The SAN Book III – October 22, 2001
iSCSI
iSCSI sends SCSI iSCSI sends SCSI commands as data, (as does Fibre Channel), but to transport the
commands as data, (as data, it uses Ethernet layers instead of the Fibre Channel layers. Additionally, it uses
does Fibre Channel), but
to transport the data, it
a new protocol called iSCSI to replace the FC-four layer that Fibre Channel uses.
uses Ethernet layers This protocol can be used on IP SANs only.
instead of the Fibre
Channel layers. Figure 248 below compares Ethernet, iSCSI, iFCP, and Fibre Channel. Note:
Protocol mapping is more of an art than a science and interpretations may vary.
Figure 248. Ethernet, iSCSI, iFCP, Fibre Channel
Fibre
Channel Ethernet iSCSI iFCP
FC-2
FC-1
FC-0
Figure 249 below illustrates how newer protocols take elements of older ones.
ISCSI takes elements of both SCSI and IP, similar to the way in which Fibre
Channel took SCSI and morphed it into Fibre Channel by adding networking
characteristics.
Figure 249. Newer Protocols Use Elements of Older Ones
Older Newer
SCSI FC
Ethernet iSCSI
Source: Salomon Smith Barney
iFCP
The difference between Similar to Fibre Channel and iSCSI, iFCP sends SCSI commands as data but to
iFCP and iSCSI is that transport the data, it uses Ethernet layers instead of Fibre Channel layers. The
iFCP keeps the FC-4
layer instead of replacing
difference between iFCP and iSCSI is that iFCP keeps the FC-4 layer instead of
it with iSCSI. replacing it with iSCSI.
250
The SAN Book III – October 22, 2001
FCIP
The method of taking a complete frame in a different protocol and wrapping it in yet
another protocol is known as tunneling. FCIP is an open, standard way of tunneling
Fibre Channel through IP. Proprietary methods of tunneling also exist.
FCIP takes a Fibre FCIP is a combination of Ethernet and Fibre Channel. It takes a Fibre Channel
Channel frame of frame of information and then adds Ethernet layers to it. This results in more data
information and then
adds Ethernet layers to it.
needed to be transmitted for a given file (see Figure 251 below) than in either
protocol by itself because after adding all of the Fibre Channel layers, it then must
add several IP layers of instructions.
251
The SAN Book III – October 22, 2001
FC FCIP Ethernet
FC-4 FC-4
FC-3 FC-3
FC-2 FC-2
IP IP
252
The SAN Book III – October 22, 2001
253
The SAN Book III – October 22, 2001
Peak of inflated
expectations
Plateau of
productivity
Attention Slope of
enlightenment
IP Storage
Trough of
Fibre
disillusionment
Channel
Trigger Fails to establish
compelling end-user
value
InfiniBand
Time
Source: Salomon Smith Barney for protocol positioning. Dataquest for Technology “Hype” Cycle.
In the early days of Fibre Channel people believed it was going to do everything
including the file sharing. We believe IP storage will go through a similar evolution.
In its infancy, IP storage is currently being hailed as the all-encompassing
replacement to Fibre Channel, and is rapidly attracting attention. We believe it will
eventually occupy a solid segment of the market and co-exist with Fibre Channel.
According to industry pundit Mike Gluck, “IP storage will happen slower than the
‘hype’ but faster than Fibre Channel evolved.”
254
The SAN Book III – October 22, 2001
The PCI and PCI-X Bus architectures create an inefficient bottleneck as data gets
congested as it passes through the bus. InfiniBand overcomes this by implementing
a switched fabric chip architecture to direct and control the flow of data; thereby,
eliminating the PCI bottleneck. InfiniBand proposes to offer greater connectivity
255
The SAN Book III – October 22, 2001
with increased performance by enabling data to flow into and out of devices more
efficiently.
256
The SAN Book III – October 22, 2001
Source: Agilent
3GIO and HyperTransport are two other I/O technologies that are being discussed as
replacements for PCI, but are beyond the scope of this report.
257
The SAN Book III – October 22, 2001
What Is InfiniBand?
InfiniBand proposes to bring these I/O connections “outside” the servers, so
computing power can be upgraded or added in a simple, efficient manner. In
addition, I/O expansion takes place across a switched fabric environment,
eliminating slot limitations of PCI and PCI-X.
IDC projects servers with InfiniBand capability will grow to 6,720,296 in 2005 from
3,806 units in 2001, representing a 346.1% CAGR (Figure 256 below). Although
we believe this is a bit optimistic, it drives home the point that InfiniBand is coming
and is going to be a big market opportunity.
InfiniBand is a complete Figure 256. InfiniBand-Enabled Server Shipments
new architecture; the
transition won’t be easy
or quick. Units
7,000,000
4,000,000
3,000,000
2,000,000
1,000,000
0
2001 2002 2003 2004 2005
Source: IDC (May, 2001)
But what exactly is it? Is it a new bus or a protocol? Does it reside in the server or
outside? InfiniBand is an architecture based on a switched-fabric of serial I/O
channels; InfiniBand is so broad-reaching that it redefines existing I/O systems,
connections, cables, physical layer and high-level software protocols. Various I/O
subsystems such as Ethernet, Fibre Channel, SCSI, or even inter-processor
communication will be able to link to an InifinBand fabric. But the transition won’t
be easy or quick nor is it guaranteed to happen. Systems supporting InfiniBand will
likely first be introduced in 2002, even under the most optimistic estimates.
Think Outside the Box, Literally
InfiniBand moves the I/O A significant transition to be brought about by InfiniBand is the decoupling of the
subsystem out of the I/O subsystem from the server’s microprocessor/memory complex. Today’s server
server.
I/O subsystem consists of shared PCI bus, linked directly to the system memory and
CPU. InfiniBand architecture would replace the shared bus with a high-speed serial
link, with initial specification of 2.5 Gbps. The I/O subsystem would essentially
reside “outside” the server, within a message-passing, IP-based switched fabric.
258
The SAN Book III – October 22, 2001
IB IB
Switch Switch
Storage Subsystems
TCA
IB
Network Router
TCA
Interface
Other IB
Network
Ethernet TCA subnets,
Interface
WANs
Fibre Channel
Source: Lane15
259
The SAN Book III – October 22, 2001
CPU CPU
High-end
CPU CPU
Memory Controller
IB IB IB
HCA HCA HCA Multiple
Infiniband
Ports
260
The SAN Book III – October 22, 2001
Low-end Mid-range
CPU CPU
CPU
Proprietary Internal Data Bus PCI Bus/
Proprietary Bus
Memory Controller
IB Memory Controller
HCA
Multiple IB IB IB
Infiniband HCA HCA HCA Multiple
Ports Infiniband
Ports
261
The SAN Book III – October 22, 2001
Virtual
Standard Interface
Application Application
Operating System
Driver
HBA HBA
Using VI, the OS and the driver steps are bypassed, resulting in the data moving
more efficiently through the computer and out onto the network. In effect, VI
allows the data to take a shortcut from the application to the HBA. This is also
referred to as Direct Memory Access (DMA) since the data takes a direct path from
the main memory (application) to device memory (an HBA in this case) versus
going an indirect path by passing through additional steps. Note: In practice, the
driver is not completely eliminated but replaced by a thinner one with less
instructions.
262
The SAN Book III – October 22, 2001
Application
Presentation VI
Session
Transport
Network Fibre Channel,
Ethernet,
Datalink InfiniBand, etc.
Physical
Source: Salomon Smith Barney
263
The SAN Book III – October 22, 2001
Using VI over InfiniBand, servers can be aggregated much the same way that
storage utilizes RAID (Redundant Array of Independent Disks). For the server, we
will refer to this as RAIM (Redundant Array of Independent Microprocessors) See
Figure 264 below.
Figure 264. Potentially RAID and RAIM Architectures Could Look Very Similar
N
L AN
WA
N
LAN
WA
RAID Controller SAN SAN RAIM Controller
Fibre Channel
Infiniband
JBOD Microprocessor Farm (1U)
264
The SAN Book III – October 22, 2001
WAN
LAN
SAN
IB
In an InfiniBand World...
Still in the concept stage, As part of the press surrounding InfiniBand, there has been a lot of noise about
InfiniBand won’t be InfiniBand eliminating the need for Fibre Channel and SANs. While it is true that
ready for deployment for
many years.
someday InfiniBand may challenge Fibre Channel for SANs given that the first-
generation InfiniBand specification has only recently been completed, we believe
the technology is still in its infancy and will take time to mature as a SAN
technology. Just as Fibre Channel was incubated for more than five years,
InfiniBand will also require several iterations before it’s ready for General
Availability deployment.
The Fibre Channel HBA Let’s fast-forward to 2003, when InfiniBand products are probably just starting to
suppliers will likely also gain momentum. At this time, Fibre Channel–based SANs and 10 Gb Ethernet
offer InfiniBand HCAs.
LANs will already be deep into deployment and often working in cohabitation.
InfiniBand will likely be deployed at the core of a data center, and connect to
existing LANs, WANs, and SANs. The question is then, which vendors will bring
InfiniBand products to market? Indeed, members of the server OEM, LAN, WAN,
and SAN communities have all been involved. We believe that given their close
working relationship with server vendors, our Host Bus Adapter suppliers, the same
ones offering Fibre Channel HBAs today, will be in the forefront of offering
InfiniBand host and target devices.
In short, we believe that everyone wants to use, and will use, InfiniBand.
265
The SAN Book III – October 22, 2001
IBM
On February 21, 2001, IBM announced the IP Storage 200i, a NAS appliance that
uses the iSCSI protocol. IBM uses the same version of iSCSI (as Cisco) for its
product. Cisco and IBM jointly submitted the use of SCSI over TCP/IP to the
Internet Engineering Task Force (IETF), referred to as iSCSI.
Brocade
We believe the Cisco/Brocade partnership provides a very high level of validation
for the benefits and importance of Fibre Channel networking technology.
On June 13, 2000, Brocade announced a partnership with Cisco to interconnect
Cisco’s Catalyst 6000 product family of high performance multilayer switches with
Brocade’s Silkworm product family of Fibre Channel switches. The interconnection
will occur through the integration of a Fibre Channel–based interface card, code-
266
The SAN Book III – October 22, 2001
named Redhead, based upon Brocade’s e-port technology and located within a Cisco
switch, providing switch-to-switch connectivity for remote Fibre Channel SANs.
Cisco will pay Brocade a “license fee” on a per-port-shipped basis.
The combined solution is based on encapsulating the Fibre Channel protocol over
TCP stacked with IP through Ethernet to achieve non-blocking OC48 performance
(which is about 248 MBps, Fibre Channel currently runs at 200 MBps). Cisco will
continue to increase the speed of transmission with future generations of the
interface.
Cisco will also offer Cisco will also offer DWDM connectivity for Brocade to traverse Metropolitan Area
DWDM connectivity for Networks (MANs) and Wide Area Networks (WANs). (Note: ONI Systems is also
Brocade to traverse
Metropolitan Area
a Brocade partner and provides Brocade with DWDM. In addition, Brocade helped
Networks (MANs) and fund ONI before it went public.)
Wide Area Networks
(WANs). NuSpeed
On July 27, 2000, Cisco announced a definitive agreement to acquire NuSpeed, a
privately held company focused on IP based storage networking, for $450 million.
(The transaction closed in the quarter ended October, 2000.)
On April 9, 2001, Cisco announced the SN 5420 Storage Router, which utilizes
Cisco’s version of iSCSI. QLogic’s ISP2200A single-chip processor is being used
in the router. Some of the technology of the router came from the NuSpeed
acquisition.
Figure 266. IP SAN Router
Fibre
iSCSI Channel
Disk
Array
Router
Tape
Library
Source: Salomon Smith Barney
267
The SAN Book III – October 22, 2001
We believe this is an important announcement for iSCSI versus iFCP and FCIP. In
fact, Cisco’s announcement to acquire NuSpeed could have been an effort to stack
the deck in favor of Cisco’s iSCSI as the preferred IP storage networking standard.
This may seem to conflict with Cisco’s announcement that it was going to license
Brocade’s Fibre Channel technology to interconnect SANs; in fact, it does a bit in
one respect. With NuSpeed’s solutions (which, similar to Brocade’s technology, can
be embedded into a Cisco router and/or switch), Cisco will be able to interconnect
any Fibre Channel equipment through an IP network. With the Brocade technology,
Cisco will only be able to interconnect Fibre Channel equipment that conforms to
Brocade’s e-port standard (e-port is an inter-switch linking port). The
Cisco/NuSpeed router is more oriented to Fibre Channel node connectivity (i.e.,
storage, tape, etc.) versus Cisco’s other product, which connects to a Fibre Channel
SAN for long-distance connectivity.
Although we agree that efforts to develop IP based storage networks are gaining
momentum, we believe that IP-based storage networks will mainly service the lower
end and periphery of the market and will take at least two to three years to gain
traction in the enterprise. In the meantime, we believe that Fibre Channel will
continue to gain momentum as well.
Note: We do not expect Brocade to directly generate significant revenues from its
Cisco relationship. We do expect that Cisco will continue to develop and implement
Brocade’s e-port connectivity since it has been accepted as the Fibre Channel
interconnect standard by the American National Standards Institute (ANSI).
Separately, Brocade’s Fabric Shortest Path First (FSPF) routing protocol, a method
of routing data between various vendors’ Fibre Channel switches, was accepted as
the industry standard by the ANSI Technical Committee T11. (This is different, but
complementary, to Brocade’s e-port technology, which is the physical connection.)
EMC Q&A:
Q: Will Ethernet be a viable server-to-storage interconnect?
A: “Yes, but not soon.” -EMC
Q: When will Ethernet technology be ready for prime time as a server-to-storage
interconnect?
A: “Two to five years. But Ethernet will not replace Fibre Channel. It will be
bridged into existing SANs.” -EMC
268
The SAN Book III – October 22, 2001
269
The SAN Book III – October 22, 2001
Key Considerations
1)Relatively low cost of
Client 1
Server 1
management
EMC
2)Fault tolerant
Symmetrix 1 3)Data transmission speeds
Client 2
LAN EMC
Celerra 1 are somewhat constrained
by the LAN
EMC
Symmetrix 2 4)Heterogeneous
Client 3 N
SA connectivity
EMC
5)Increased scalability
Celerra 2 EMC
Client 4 Traditional Symmetrix 3
NAS
Appliance
Tape
Library 2
Figure 268. SAN, NAS, and DAS (Direct Attached Storage) Convergence with EMC’s Celerra
WAN/MAN/LAN
Tape
Library
Application
Server 1 EMC
Client 1 Sym 1
EMC
Client 2
LAN SAN Symmetrix 2
Application
Server 2
Client 3 EMC
Symmetrix 2
EMC Celerra
NAS Tape
Appliance NAS
Appliance 1
270
The SAN Book III – October 22, 2001
EMC
Client 2
LAN SAN Symmetrix 2
271
The SAN Book III – October 22, 2001
272
The SAN Book III – October 22, 2001
Coverage Universe
273
The SAN Book III – October 22, 2001
274
The SAN Book III – October 22, 2001
7
Coverage Universe
275
The SAN Book III – October 22, 2001
276
The SAN Book III – October 22, 2001
In second quarter 2001, enterprise hard disk drives accounted for over one-third of
suspension production, desktop drives accounted for about 50%, and mobile hard
drives accounted for the remainder.
Hutchinson works closely with its customers to develop products that improve disk
drive performance and reduce manufacturing cost.
277
The SAN Book III – October 22, 2001
OEMs such as EMC, Compaq, and StorageTek. Most of JNI’s success to date has
been from connecting Sun servers to storage subsystems from EMC, Compaq,
StorageTek, Hitachi Data Systems, and others. Given the stability and popularity of
Sun’s Solaris platform, JNI has benefited tremendously from its growth. The
company is now building on its Solaris success, and expanding into other UNIX-
based architectures such as HP-UX and IBM’s AIX.
278
The SAN Book III – October 22, 2001
information (nine terabytes in the near future). NetApp Filers feature built-in RAID,
clustered failover and redundant components for increased reliability.
279
The SAN Book III – October 22, 2001
280
The SAN Book III – October 22, 2001
Private Companies
281
The SAN Book III – October 22, 2001
282
The SAN Book III – October 22, 2001
7
Private Companies
Note: Companies are listed in alphabetical order. The length of individual
company profiles is based on the availability of information and not on preference.
For example, companies we have previously published on typically have longer
profiles, while companies in stealth mode have shorter ones.
283
The SAN Book III – October 22, 2001
RAID controllers and 3DM Remote Storage Management software, creates one of
the most promising emerging low-cost storage solutions. The 3ware solution allows
the transmission of block-level SCSI data over Gigabit Ethernet LANs or subnet
SANs, enabling customers to leverage existing investments in network
infrastructures, reduce management complexity, significantly lower overall costs,
and potentially eliminate the need for Fibre Channel IT personnel. The 3ware NSU
is available in configurations from 240GB to 600GB and supports Windows 2000,
Windows NT, MacOS, Solaris, and Linux platforms.
3ware’s Escalade controllers, which are packaged with the NSU, are based on the
company’s DiskSwitch packet-switching technology. 3ware developed its
DiskSwitch Architecture to decrease latency and increase performance. The
DiskSwitch storage controller architecture applies the principles of packet-switched
networking to subsystem I/O. Each disk drive has fully dedicated bandwidth,
eliminating bus arbitration delays and other limitations of shared-bus architectures.
Low-cost UltraATA drives are used by the company in place of SCSI or Fibre
Channel drives, and the company claims this is without incurring any performance
penalty. Escalade controllers are available in two-, four-, and eight-port models and
can be sized to fit any configuration from a 1U rack-mounted system to a high-
capacity, multi-controller system.
3ware’s 3DM Remote Storage Management software displays drive and RAID array
status, allows remote configuration of new arrays, and enables hot swapping of
failed drives for on-the-fly rebuilding. 3DM’s Web-based user interface provides a
common look and feel for all platforms, and, because it is browser-based, customers
can manage their storage from any point of network access without installing
software on the remote system.
284
The SAN Book III – October 22, 2001
Alacritech intends to offer iSCSI adapters in late 2001. Additionally, the company
has started developing a 10 Gb Ethernet solution. Depending upon completion of
the standard, it plans to offer this product in 2003.
Figure 270. 100x4 Quad-Port Server Accelerator
In June, 2001, Alacritech announced its 100 Series Server Accelerators support MS
Windows 2000 and Redhat Linux operating systems. Alacritech’s founder, Mr.
Larry Boucher (inventor of the SCSI interface) has previously founded two
successful technology companies, Adaptec (ADPT) and Auspex Systems (ASPX).
The company is privately held and has completed two rounds of funding totaling
$22.4 million from Benchmark Capital, Institutional Venture Partners, Alloy
Ventures, and Berkeley International.
285
The SAN Book III – October 22, 2001
integrators, value-added resellers, and SCSI computer end users, as well as network
equipment suppliers.
Ancot was founded in 1987 by President Jan Dedek, to provide SCSI test and
development equipment to the microelectronics industry. Since then, Ancot’s
products have evolved, keeping pace with the growth and changes of the SCSI bus
protocol. Ancot shipped the industry’s first Fibre Channel analyzer in 1994 and
continues to develop SCSI and Fibre Channel products. Combining SCSI and Fibre
Channel technologies, Ancot now offers a more complete line of analyzer and
extender solutions.
For its upper tier customers, Ancot offered the industry’s first 2 Gb protocol
analyzer for Fibre Channel systems. Ancot also offers a server version Fibre
Channel Analyzer that supports 1.06 Gbps data rates, yet cuts in half the typical cost
for advanced Fibre Channel instrumentation.
➤ FCAccess 2000 Fibre Channel Analyzer: supports both 1.06 Gbps and 2.12
Gbps data transfer speeds and is a leading network-attached, client-server Fibre
Channel Analyzer.
Figure 272. Ancot — FCAccess 2000 Fibre Channel Analyzer
➤ The Ultra2160 and Ultra2160/Lite SCSI Bus Analyzers: premier tools for
debugging and optimizing SCSI devices and systems.
Figure 273. Ancot — The Ultra2160 SCSI Bus Analyzer
The company’s world headquarters is located in Menlo Park, California. Sales and
support for U.S. customers are provided from this location, while international
markets are reached through its worldwide distributor network.
286
The SAN Book III – October 22, 2001
287
The SAN Book III – October 22, 2001
288
The SAN Book III – October 22, 2001
The ATTO ExpressPCI FC 3305 brings 2 Gb throughput to server cluster and SAN
environments. The FC 3305, with a fixed copper HSSDC Fibre Channel interface,
allows for high-speed copper connections. Designed to support next-generation
SAN requirements, the FC 3305 is ideal for high-speed random or sequential types
of data-intensive environments.
289
The SAN Book III – October 22, 2001
The ATTO FibreCenter 2100R serves as a high-speed, central connection point for
data sharing in Enterprise Fibre Channel Arbitrated Loop (FC-AL) configurations.
The addition of a rack-mounted Fibre Channel hub in the FC Rack System gives
Enterprise users countless options for configuring, upgrading, and expanding Fibre
Channel Storage Area Networks (SANs). This five-port hub provides full duplex
290
The SAN Book III – October 22, 2001
1.0625 Gb performance and delivers 200 MBps transfer rates for I/O and data-
intensive SAN environments.
Figure 278. ATTO FibreCenter 2100R
291
The SAN Book III – October 22, 2001
292
The SAN Book III – October 22, 2001
➤ Ultra-SCSI modular RAID subsystem with failover, UPS, Battery Backup and
Caching. Capacities from 34 GB to 276 GB of usable disk space.
Figure 280. Baydel’s O/ESR RAIDER-5
293
The SAN Book III – October 22, 2001
Figure 281. An Example RAIDMON Software Screen. Error Notification: Failed Disk
Baydel works closely with value-added resellers that can specialize in particular
platform or application environments and provide high-quality integrated solutions
and services. Baydel offers support groups in Leatherhead and San Jose, direct
service in many North American cities, and trained service partners, including IBM,
SUN, Digital, Silicon Graphics, and Motorola in certain geographic territories.
294
The SAN Book III – October 22, 2001
On May 7, 2001, BlueArc completed a $72 million round of funding, bringing total
funding since the company’s founding to over $100 million. This round will be used
to increase company growth. Investors from this round include Weston Presidio,
Celtic House International, Apax Partners, and Patricof & Co., Dell Ventures,
EUROQUBE, NY State Retirement Co., J.P. Morgan Partners, and Fort
Washington Capital.
295
The SAN Book III – October 22, 2001
296
The SAN Book III – October 22, 2001
297
The SAN Book III – October 22, 2001
298
The SAN Book III – October 22, 2001
299
The SAN Book III – October 22, 2001
Figure 287. Chaparral Network Storage — FS2620: 2Gbps Fibre Channel-to-Ultra160 SCSI Storage
Router
The Company sells its products to OEMs, including Eurologic Systems, MicroNet
Technology, Inc., Qualstar Corporation, Quantum Corporation/ATL Products, Inc.,
Trimm Technologies, Inc., and Xyratex International Ltd., as well as to distribution
partners, including Arrow Electronics, Inc., Bell Microproducts, Inc., and CONSAN.
In late June, 2001, Chaparral announced the appointment of Mr. Victor Perez to the
position of chief executive officer. Mr. Perez has been Chaparral’s president and
COO since joining the company in December, 2000, and has served on Chaparral’s
board of directors since August, 2000. He will retain his president and COO roles.
Mr. Gary Allison, who had been Chaparral’s CEO since the company’s inception in
1998, will continue to serve as chairman of Chaparral’s board of directors.
300
The SAN Book III – October 22, 2001
The company was originally founded in 1986 to develop and market SCSI host
adapters for Digital Equipment Corp. (DEC) computer systems. Today, CMD is a
leader in Fibre Channel and SCSI RAID controller solutions for UNIX, open
systems and PC marketplaces and is a leading supplier of IDE/Ultra ATA and
Universal Serial Bus (USB) semiconductors for the PC and embedded marketplaces.
CMD continues to invest in the computer peripheral interface and data-storage
technology markets
CMD provides data-storage technology for a wide range of open systems, high-
performance workstations, LANs, and PCs. CMD’s storage products are marketed
through system integrators, VARs and OEMs — many of the leading providers of
storage solutions. From CMD’s early success in the DEC market, many product
initiatives evolved. In 1991, CMD made the decision to add RAID to its product
line and to expand into the PC, UNIX and open-system marketplaces. Two years
later, CMD released the CRD-5000, its first RAID product.
The CMD Titan CRA-7280 RAID controller is a high-end SAN RAID controller
providing dual-redundant, hot swappable, active/active or active/passive RAID
controllers, power supplies, and cooling fans, dual Fibre Channel host interfaces,
and eight LVD Ultra2 disk channels. It provides true enterprise level storage
capacities and performance and comes bundled with an easy-to-use network-based
GUI RAID management utility.
Figure 289. CMD Technology — CMD Titan CRA-7280 RAID controller
CMD also develops, manufactures, and sells high-performance ASIC and board-
level products. CMD’s IDE and Ultra ATA chips can be found in computer
systems, including those by ACER, AST, Compaq, Dell, Digital, Gateway, Hewlett-
Packard, and others. CMD also provides a wide range of complementary
software/firmware products and services that are designed to enhance the total
product solution.
301
The SAN Book III – October 22, 2001
302
The SAN Book III – October 22, 2001
303
The SAN Book III – October 22, 2001
304
The SAN Book III – October 22, 2001
In June, 2001, JNI’s entire family of FibreStar Fibre Channel HBAs was verified
compatible and interoperable with DataCore’s SANsymphony software.
Additionally, the company announced that SANsymphony was selected as one of
the top five storage products of the year in InfoWorld’s, Reader’s Choice Award
issue.
The SAN DataDirector provides continuous access to your data and integrates easily
into multiplatform environments.
305
The SAN Book III – October 22, 2001
306
The SAN Book III – October 22, 2001
software base provided in its SAN Operating System and supported in its SAN Data
Manager to enable a centrally-managed control point from which its entire SAN
solution can be integrated. As such, DataDirect is able to significantly differentiate
its hardware offering and ensure it is positioned for future growth as new
capabilities, systems, capacity, and applications are added into its customer’s SANs.
DataDirect Networks also offers SAN RAID Systems, a family of products that
provide full redundancy, speed, high data availability, capacity, and scalability to
deliver terabytes of data in a single system from 200 MBps to 35 Gbps.
Figure 293. DataDirect Networks’ OEM Exclusive EV-5000, Fastest Fibre Channel Network RAID System
— 185MB/Sec Sustained Throughput
Figure 294. DataDirect Networks’ EF-2000 Fibre Channel SAN Ready RAID Solution
DataDirect Networks has sold over three petabytes of storage on more than 30,000
systems. The company’s customers include Amoco, Eastman Kodak, Microsoft,
NASA, Panasonic, Pinnacle Systems, Sony, and StorageTek. DataDirect Networks
has strategic partnerships with ATTO technologies, Brocade, Emulex, Gadzoox,
JNI, Qlogic, Seagate, and Vixel.
307
The SAN Book III – October 22, 2001
308
The SAN Book III – October 22, 2001
309
The SAN Book III – October 22, 2001
Thunder
Lightning 9200
9900
310
The SAN Book III – October 22, 2001
Key Alliances
VERITAS Software
On February 14, 2001, HDS, Hitachi, Ltd., and VERITAS Software announced the
formation of a Global Storage Management Agreement, which focuses on the
coordinated integration and testing of the companies’ storage solutions, in order to
architect highly interoperable data availability configurations that ensure increased
flexibility for the end user. As part of this Global Storage Management Agreement,
HDS will offer the latest versions of a wide range of VERITAS Software products
and professional services to its customers around the globe. The companies will
also unite in joint product integration and certification efforts.
Hewlett-Packard
On the OEM side, one of HDS’s top server partners is Hewlett-Packard. On April
28, 1999, HDS announced that under a three-year OEM joint technology agreement,
Hewlett-Packard will enhance its broad range of SAN and storage-management
products with HDS’s products and technology (this agreement was recently renewed
for another three years). More specifically, Hewlett-Packard will OEM high-end
subsystem technology from HDS (the Lightning 9900) and cooperatively enhance
products with Hewlett-Packard-contributed firmware and fibre channel interconnects
for improved open systems performance. Hewlett-Packard will also leverage HDS’s
311
The SAN Book III – October 22, 2001
Sun Microsystems
On August 8, 2001, Sun and Hitachi Data Systems (HDS) announced a three-year
partnership agreement. Under the terms of the agreement, Sun will include HDS’s
high-end Lightning 9900 in its new Sun StorEdge 9900 series. In addition to the
subsystem itself, the Sun StorEdge 9900 also includes access to three jointly-staffed
support centers and certification in SunPlex and SunCluster environments. Other
terms of the agreement include cross-licensing of software and collaboration on
future software products.
We believe HDS is clearly focused on expanding its breadth of storage OEM and
reseller partnerships. Based on HDS’s success, we believe these partnerships could
make a lot of sense in the near term. We expect revenues will grow as a result of
these partnerships and HDS’s stated intention to grow channel sales twice as fast as
direct sales going forward. HDS expects indirect sales to account for 50% of total
sales within two years.
312
The SAN Book III – October 22, 2001
The company’s Satellite Fibre Channel analyzer offering includes the portable IFC-
30 and IFC-40 Fibre Channel analyzers that monitor two channels of data at transfer
rates of 1 or 2 Gbps, respectively. It also includes the IFC-3016 and IFC-4016
desktop models that simultaneously monitor up to 16 channels at the same
respective speeds — for optimum efficiency and control at an extremely low cost
per channel.
Figure 298. I-TECH —Satellite IFC-4016 Fibre Channel Analyzer
313
The SAN Book III – October 22, 2001
314
The SAN Book III – October 22, 2001
315
The SAN Book III – October 22, 2001
Lane15 Software, named for the “virtual lane” within the InfiniBand architecture
that is reserved for management systems, is a developer of vendor-neutral fabric
management solutions that can be readily adopted by any InfiniBand technology
provider, including chip manufacturers, server vendors, networking device
providers, and I/O vendors. In early July, 2001, Lane15 successfully demonstrated
its Fabric Manager technology at the IBTA Spring Developers’ Conference. With
its strong backing and demonstrated product, Lane15 Software is actively advancing
this next generation switched fabric I/O architecture.
316
The SAN Book III – October 22, 2001
Enterprise Directory
Console Console
Web GUI
http(s)
http(s) xml
xml
Lane15 Fabric Manager LDAP or
Active
Directory
Policy Engine
extensions
Mechanisms
Switch IOCTLR
TCA
HCA
extensions extensions extensions
Lane15 Software views the adoption of InfiniBand to start in 2001 with InfiniBand
silicon available (see partnerships below), 2002 for initial production systems, 2003
for early mainstream adopters and large scale deployment in late 2003 to early 2004.
Banding Together
Lane15 Software and Banderacom, a fabless InfiniBand semiconductor company,
have formed a partnership to deliver integrated silicon, software, and development
tools to InfiniBand equipment manufacturers. Banderacom will provide the silicon,
and Lane15 will supply the software. Lane15 and Banderacom along with
WindRiver Systems, a provider of software and services for connected smart
devices, demonstrated a prototype of the industry’s first InfiniBand Ethernet router
target channel adapter at the Intel Developer Forum crossing another key milestone
in the development of InfiniBand products. Note: Lane15 is also working with other
InfiniBand silicon suppliers such as Mellanox and Intel.
Funding History
Lane15 Software was founded in April, 2000, by Alisa Nessler, a Venture Fellow at
AV Labs. Based in Austin, Texas, the company was incubated within AV Labs after
receiving initial funding of $750,000. Lane 15 has raised approximately $9 million
in two separate rounds from a broad range of investors including Intel, Compaq,
Dell, Austin Ventures, AV Labs, and Lightspeed Venture Partners.
317
The SAN Book III – October 22, 2001
approaches that rely on a single network connection to a storage volume, the NUS
architecture avoids CPU bottlenecks and increases flexibility by mounting modules
in parallel on the network. NUS uses the network itself as the storage bus which lets
customers choose how to organize their storage volumes across Ethernet or gigabit
Ethernet networks to emphasize performance, redundancy, or both. Striping and
mirroring are possible within, and among, LeftHand’s devices — regardless of
geographical distance. The overall performance is similar to that of a Fibre-Channel
SAN, on standard Ethernet networks. The NUS architecture combines the
performance of a SAN with the ease-of-use of NAS in a stand-alone, integrated
network storage module. The company is expected to begin product shipment in
third quarter 2001.
LeftHand Networks recently received follow-on funding for its Series A round,
bringing the total to $13 million. Key investors include Boulder Ventures, Sequel
Venture Partners, and Vista Ventures I.
318
The SAN Book III – October 22, 2001
form of digital content, from e-mails to audio to still images and movies. We
believe application-oriented services aimed at vertical markets offer higher value-
add and likely command higher gross margins.
The company has raised over $60 million to date from investors including Great Hill
Equity Partners, Providence Equity Partners, First Union Corporation, J.P. Morgan,
and EMC. StorageTek still owns approximately 15% of the company.
Source: Mellanox
319
The SAN Book III – October 22, 2001
320
The SAN Book III – October 22, 2001
➤ The IPS 2000 switch includes up to six ports: two Gigabit Ethernet connections
and four Ultra2 SCSI (80 MBps) or Wide Ultra SCSI (40 MBps) ports.
321
The SAN Book III – October 22, 2001
➤ The IPS 3000 switch has eight MultiService Interface ports, enabling users to
mix Fibre Channel and Gigabit Ethernet connections.
Figure 304. Nishan Systems’ IPS 3000 Switch
Nishan Systems also offers the SANvergence Management Suite, which is based on
SNMP and Java, for managing the SoIP devices,. The software enables users to
manage an IP storage fabric consisting of Nishan’s switches and gateways. The
graphical interface of these tools provides an intuitive display and modification of
configuration data — simplifying the setup of switches and gateways.
Figure 305. Nishan Systems’ SANvergence Management Suite
Earlier this year, Nishan Systems formed a broad strategic agreement with JNI
Corporation covering joint research, development, and marketing opportunities for
the integration of Storage over Internet Protocol (SoIP), including iSCSI, solutions
with Fibre Channel–based SANs in Metropolitan Area and Wide Area Networks
(MANs and WANs). The companies have entered into a “first-to-market”
agreement under which the two companies will jointly market and sell their
products.
Recent news includes Veritas interoperability certification of Veritas Backup Exec
with Nishan’s IP Storage Switches as well as a joint announcement with IBM and
Intel that software for the IP Storage protocol, Internet Storage Name Service
322
The SAN Book III – October 22, 2001
(iSNS), has been released as open source code. In addition, Nishan announced that
Intel will also support the iSNS protocol by enabling the Intel PRO/1000 IP Storage
Adapter, which will be available later this year, to utilize iSNS information.
323
The SAN Book III – October 22, 2001
324
The SAN Book III – October 22, 2001
number of drives (spindles) become crucial for immediate I/O access, according to
the company.
Figure 307. The NexStor 3250S — Dual RAID Controller SCSI Storage System
The NexStor 802S and 1202S provide unique SCSI JBOD storage solutions that fit
in a compact space. The NexStor 802S accommodates up to eight 1.0-inch or 1.6-
inch Ultra160 SCSI disk drives, including the 180 GB drive, mounted horizontally
in a 3.5-inch (2U) enclosure, while the NexStor 1202S accommodates up to 12 one-
inch drives in the same form factor. NStor’s design supports 7,200 RPM, 10,000
RPM, and 15,000 RPM disk drives, and allows an easy upgrade path to an
active/active SCSI or Fibre Channel host-attach RAID solution.
Figure 308. The NexStor 1202S — JBOD SCSI Storage System
nStor has ensured that all of its products are SAN-ready and has developed a number
of products specifically for the SAN market, including its GigaRAID/FC family
which offers Fibre Channel host connectivity with performance up to 200 MBps.
nStor sells its products through a worldwide network of OEMs, and distributors.
The nStor customer base ranges from small businesses to Fortune 1000 firms and
government organizations, including Motorola, Yellow Pages, Whirlpool, Bank of
America, Trans America, Union Pacific, Prudential Insurance, The Trane company,
Intergraph, Warner Brothers, EDS, Lucas Film Ltd., Dream Works, the U.S. Navy,
and the U.S. Air Force.
Additionally, nStor’s AdminiStor Plus RAID configuration and storage management
software suite is fully integrated for use with nStor storage solutions. AdminiStor’s
advanced capabilities give customers the ability to configure, monitor, and manage
their nStor storage resources via Web browsers on a LAN/SAN network or over the
Internet.
325
The SAN Book III – October 22, 2001
Recent news includes an announcement that nStor had received an offer from a
private investor to purchase approximately 45% interest in the company for $12
million. According to the press release, the offer requires the company to cause all
of its preferred stock to be converted into common stock, as well as other conditions.
Also in the press release, management stated that the company will evaluate the
offer, confer with the holders of preferred stock and respond to the proposal on or
before a late July, 2001 deadline provided for in the offer.
The IBgate 1000 is the first of a new series of InfiniBand Architecture (IBA)-to-
Gigabit Ethernet (GbE) gateways available from OmegaBand. It provides a
standards-based IBA-GbE interface that combines performance with networking
326
The SAN Book III – October 22, 2001
327
The SAN Book III – October 22, 2001
methods (block and file services), as well as supports a wide range of industry-
proven storage subsystems, under centralized and secure management.
Figure 311. The Pirus Storage Utility Switch
The Pirus solution addresses key storage issues facing large enterprise customers by:
➤ unifying file (NAS) and traditional block (SAN) components, IP with Ethernet
and Fibre Channel;
➤ securely sharing storage network resources among diverse communities of users
through the Pirus Secure Virtual Storage Domains;
➤ scaling new connections, such as new servers or storage resources, and enabling
additional processing power or capacity on demand;
➤ ensuring integration with the installed base of existing equipment, including
Fibre Channel fabric and director-class switches;
➤ bridging and converging network environments to enable integrated
management, efficient network utilization and greater connectivity across the
LAN, WAN, and MAN.
As the evolution of storage networking is an ongoing process, Pirus seeks to
preserve existing investments in storage systems while providing a smooth
migration path to next-generation storage architectures.
In March, 2001, Pirus acquired Blue Spruce Networks, a storage software
engineering firm located in Wichita, Kansas. Financial details of the acquisition
were not disclosed. Blue Spruce Networks will contribute its expertise in the areas
of storage architecture design, fibre channel, storage software development, and OS
platform integration — which complement Pirus’s own storage and networking
experience. Blue Spruce Networks will remain in Wichita, Kansas, serving as a
Pirus branch office, and report directly to Pirus’s vice president, engineering.
Additionally, the same month, Pirus announced that it closed over $27 million
through its Series B financing round, bringing total funding to over $45 million.
Strategic investors in this round of funding include StorageNetworks and Veritas. In
September, 2001, Pirus received an additional $9.5 million in funding from JAFO
328
The SAN Book III – October 22, 2001
Ventures. The new funding is targeted for the continued expansion of the
company’s research and development, marketing, and sales efforts.
The company is currently involved in many industry standard initiatives such as
IETF IPS Group (IP Storage), SNIA IP Storage Forum, ANSI T11, and FCIA. To
date, Pirus has raised over $55 million from strategic investors including
StorageNetworks and Veritas Software as well as financial investors such as
BlueStream Ventures, Bessemer Venture Partners, Charles River Venture Partners,
Comdisco Ventures, GATX Ventures, JAFCO, Morgan Keegan, and Silicon Valley
Bank.
329
The SAN Book III – October 22, 2001
Features include:
➤ comprehensive wellness management for high node count environments
➤ granular monitoring of server, blade, and application health
➤ any-to-any configurations with easy “Distributed Management Console”
interface
➤ continuous processing during multiple hardware or software failures
➤ starting, stopping, and moving applications around the entire server environment
➤ efficiency improvement of load balancing efforts
Figure 313. PolyServe Matrix Server
330
The SAN Book III – October 22, 2001
For enterprise product and service solutions, PowerQuest offers its PowerManage
Suite, which provides users a data availability solution for enterprise-wide Windows
NT and 2000 Servers. This suite allows users to quickly isolate and resolve storage
problems on a real-time basis. According to the company, the suite combines
PowerQuest VolumeManager, PowerExpert ST, and StorageMonitor, for a solution
that prevents storage problems before they occur, minimizes system downtime in the
event of a problem, and reduces the overall cost of network storage administration,
all while controlling critical data assets.
331
The SAN Book III – October 22, 2001
332
The SAN Book III – October 22, 2001
333
The SAN Book III – October 22, 2001
334
The SAN Book III – October 22, 2001
Fibre Channel, and NAS products to market, with several thousand sold in over 45
countries worldwide.
Raidtec provides the market with end-to-end Fibre Channel PCI RAID controllers
and systems. The company manufactures its products in company-owned facilities
in Atlanta, Georgia, and Cork, Ireland. The engineering team in Cork, focuses on
Fibre Channel, SAN software and RAID controller and enclosure design, and on
NAS technology through the nEngine business unit. Raidtec has developed a wide
range of SCSI and FC products covering Server attach, Network attach, and SAN.
Raidtec’s broad product offering enables the company to be a “one stop shop” for
distributors, integrators, and VAR’s RAID storage requirements. Raidtec’s value
proposition is that it delivers extra features, performance, and reliability that are on
the forefront of customer needs.
Figure 318. Raidtec’s Products
Raidtec’s product strategy is based on four core technology areas: Fibre Channel for
high performance RAID and SAN, NAS systems with embedded RAID, Fibre
Channel and LVD SCSI subsystems, and RAID power and packaging from two to
120 bays per system.
Raidtec’s nEngine business unit is focused on the development of e-commerce and
Internet infrastructure solutions, providing embedded, industrial-strength server
appliances for online commerce to take place in a secure, robust manner. Raidtec is
focused on the NAS market with its software and hardware for the workgroup and
departmental NAS server appliance segment of the market.
335
The SAN Book III – October 22, 2001
336
The SAN Book III – October 22, 2001
Rhapsody has received an initial round of funding worth $12 million and is close to
completing a second round of funding expected to be around $40–$50 million. Key
investors include Accel Partners and Sequoia Capital.
337
The SAN Book III – October 22, 2001
centralize storage capacity and management, as well as, consolidate servers into a
large computing farm over all of the industry standard network interfaces.
Sanera Systems is currently in stealth mode and has yet to announce its official
product offering. According to the company, its disruptive, high-performance
hardware and software architecture design boasts unmatched switching capacity,
multi-protocol connectivity, and unmatched performance. This enables service
providers and enterprises to deploy and manage a very large distributed storage area
network and server farms in a low latency, high bandwidth, service rich fabric.
In November, 2000, Sanera Systems secured its first round of financing of $17.5
million. Participants included Storm Ventures, CMEA Ventures, E*TRADE
Venture Capital, together with certain private investors. In April and May, 2001, the
Company received an additional $10 million financing from Western Technology
Investments and other private investors.
338
The SAN Book III – October 22, 2001
Hardware
The SANgate Systems ESA is an enterprise-class storage appliance that supports
both mainframe and open systems environments and operates with many vendors’
storage subsystems (according to the company). This intelligent platform —
specifically designed to run data storage applications — lets an enterprise increase
the capacity of its storage subsystems while increasing performance. It delivers
massive amounts of processing power, all dedicated to data storage management. Its
open architecture and comprehensive APIs allow it to be adapted to support future
storage system standards.
Software
The SANgate ESA Storage Management Software is an enterprise-class software
that performs seamless remote mirroring between any storage subsystems from any
vendors. It protects information while increasing the performance of your existing
environment.
SANgate Storage Management Software runs on the Enterprise Storage Appliance
(ESA). In addition to remote mirroring, it provides point-in-time copying and data
migration capability. SANgate Storage Management Software is a vendor-
independent software for managing data storage that supports both mainframe and
open systems environments.
339
The SAN Book III – October 22, 2001
StoragetoneOS is the core of the company’s Global Storage Intelligence (GSI) in-
house operating system that enables storage resource management, monitoring, and
customer provisioning within each storagecel. The company has 45 storagecels that
provide managed storage services to over 600 customers across a global footprint
from Germany to Japan. Sanrise provides these customers with real-time views of
their managed storage services through its secure website, mystoragetone.com. This
enables charting of mission-critical information flows which allows an enterprise to
more effectively manage its business and forecast future storage requirements.
340
The SAN Book III – October 22, 2001
Sanrise also provides a full suite of professional services to solve its customers’ IT
data management challenges. By conducting a complete storage assessment, the
Sanrise team of storage services experts is able to determine a customer’s managed
storage requirements and design a customized managed storage solution.
Sanrise has partnerships with leading technology providers including Brocade,
Hitachi, Oracle, StorageTek, Sun Microsystems, and Veritas, in addition to Internet
hosting companies such as Exodus Communications. The company has raised over
$160 million from investors including Comdisco Ventures, Crosspoint Venture
Partners, Exodus Communications, GATX Ventures, Hitachi Data Systems, Hitachi
Limited, Morgan Stanley Dean Witter, Texas Pacific Group (ACON Venture
Partners), and Veritas.
In addition, the company also offers the SVS Management System, an integrated
management tool for comprehensive management of the SL1000 IP-SAN Gateway.
Administrators can manage network performance, run statistical analysis, and
configure the SL1000 through its graphical user interface.
341
The SAN Book III – October 22, 2001
SAN Valley has released its Aspen Fibre Channel Managed Hub to the marketplace,
offering the industry’s only lifetime warranty. SAN Valley is a member of the Fibre
Channel Industry Association and the Storage Network Industry Association.
342
The SAN Book III – October 22, 2001
343
The SAN Book III – October 22, 2001
hardware and software into a storage solution that can become the foundation for
any organization’s storage utility. This convergence greatly reduces infrastructure
complexity, demands on IT resources and risk to data, while increasing data
accessibility. XIOtech products are optimized for dynamic enterprise storage
environments with compatibility for enterprise applications.
Figure 325. XIOtech Product Family
344
The SAN Book III – October 22, 2001
345
The SAN Book III – October 22, 2001
➤ Remote management — allows users to use the calc software tools remotely.
Shugart Technology is currently looking for a first round of financing of $15–$20
million.
AssuredStorage — MSS
AssuredStorage provides storage capacity and services tailored to unique customer
needs on a “pay as you grow” basis. As with other managed storage service models,
companies do not need to purchase equipment. In addition, Storability assumes the
risks and burdens associated with storage, including implementation, product
reliability, storage management automation and oversight, monitoring and
equipment service, procurement, and capital financing.
AssuredAbility — Traditional Services
AssuredAbility are Storability’s consulting services for designing, implementing,
and managing open SANs. Storability’s personnel determine requirements and
346
The SAN Book III – October 22, 2001
347
The SAN Book III – October 22, 2001
348
The SAN Book III – October 22, 2001
➤ The Tantia Enterprise Agent Suite, which allows online backup and recovery of
business applications: Oracle, IBM DB2, Microsoft SQL Server, Sybase,
Informix, SAP R/3, Lotus Domino, and Microsoft Exchange.
349
The SAN Book III – October 22, 2001
➤ Tantia High-Speed File Transfer (Tantia HFT) — high-speed file transfer using
a variety of high-speed connectivity options across enterprise platforms.
Figure 330. Tantia Technologies — RS/6000 Solution
350
The SAN Book III – October 22, 2001
351
The SAN Book III – October 22, 2001
TNS GUI-based configuration and monitoring tools make setup and administration
easy. The TNS 1600 solutions support Windows 2000 Server, Advanced Server,
and Redhat Linux, allowing easy migration to a familiar network platform,
increasing users’ comfort levels, speeding access times, and simplifying system
management. The company’s solutions use standard protocols and are compatible
with popular third-party backup and management tools.
The company’s product should appeal to a small or mid-sized business, where a
complex SAN or NAS has not yet been installed. Time N Systems’ solution could
provide an attractive alternative for companies not willing to install a rudimentary
Fibre Channel SAN configuration or high-end server.
The company has raised approximately $15 million in two rounds of fundraising.
Key investors include Austin Ventures, Convergent Investors, and Sternhill Partners.
352
The SAN Book III – October 22, 2001
353
The SAN Book III – October 22, 2001
354
The SAN Book III – October 22, 2001
connected SAN devices from a single location anywhere on the network, and we
believe this type of technology is essential for large-scale networks.
As an enterprise-level device management solution, ACM serves as a stand-alone
application or can be fully integrated within higher-level management frameworks
to provide a comprehensive SAN management solution. Additionally, ACM allows
administrators to easily create highly available network configurations by centrally
applying Path Command to specific paths to provide dynamic load balancing and
automated failover and failback capabilities.
ACM improves SAN performance, enhances end user applications, and resolves
network problems by providing remote configuration and firmware and driver
upgrades, diagnostic and troubleshooting capabilities, and performance tuning tools.
This centralizes network control and helps to reduce network administration costs, in
our opinion, essentially freeing administrators from the manual installation of
software, controller setting configuration, and troubleshooting otherwise needed to
maintain an efficient Fibre Channel network.
To date, Troika Networks has raised over $150 million from a number of financial
and strategic investors. Early investors included Draper Fisher Jurvetson and
Dynafund Ventures. Later investors include New Enterprise Associates, Winward
Ventures, Amerindo Investment Advisors, Intel Capital, and Network Appliance.
Troika Networks has partnered with Network Appliance to integrate Troika’s Zentai
Controller technology into Network Appliance’s enterprise filters. Troika’s solution
utilizing the VI (Virtual Interface) standard will enable integration of Network
Appliance storage appliances within Fibre Channel storage networking
environments, powering high-performance server-to-server and filer-to-filer data
exchange. Troika is also a member of McData’s OpenReady program for
comprehensive SAN interoperability, and on February 27, 2001, Troika announced
that its Zentai Controller had passed interoperability testing for Brocade products in
the “Brocade Fabric Aware” program.
Troika is an active member of the Storage Networking Industry Association (SNIA),
the Direct Access File System (DAFS) Collaborative, the Fibre Channel Industry
Association, the Fibre Alliance, and the InfiniBand Trade Association.
355
The SAN Book III – October 22, 2001
356
The SAN Book III – October 22, 2001
TrueSAN sells its products and services directly to end users and integrators in order
to leverage the advantages of the direct business model.
357
The SAN Book III – October 22, 2001
358
The SAN Book III – October 22, 2001
359
The SAN Book III – October 22, 2001
Tight-knit Lineup
VIEO provides end-to-end, modular software for InfiniBand management. This
greatly simplifies the software integration task for partners because the entire stack,
from chipset control software to the management interface, is available from a single
source. VIEO’s ongoing product support and maintenance allow partners to remain
focused on their core business competencies. VIEO’s current products include:
VIEO Fabric Manager, VIEO FabricView, and the VIEO Software Developer Kit.
VIEO Fabric Manager
VIEO Fabric Manager includes all fabric management functions required by the
InfiniBand architecture, plus extensions specifically for VIEO. The software
framework, on which the management applications are built, provides support for
multiple chipsets, operating systems, and execution environments. The main
components of Fabric Manager are 1) core services, 2) subnet management modules,
and 3) general services modules.
Core Services
Core services provide a software framework for writing InfiniBand transports such
as Virtual Interface, Sockets Direct, and the VIEO Management API. This
framework is independent of the specific InfiniBand chipset, operating system, and
execution environment and exists at all levels of an operating system stack. All
VIEO provided applications are written using this framework.
360
The SAN Book III – October 22, 2001
Agents Managers
Management API
Common Services
Core Services Channel O/S Services O/S Bypass
Abstraction Adaptation Hardware
Layer Layer Library
361
The SAN Book III – October 22, 2001
resources to acquire, customize, and deploy these products can continue to use their
familiar human interfaces and semantics, thereby greatly simplifying the InfiniBand
management task and reducing cost.
FabricView gives software developers a convenient and stable programming
interface to the VIEO Fabric Manager. This interface provides two-way
communications allowing InfiniBand fabric management information to be passed to
the external management application for display and monitoring, and for
management commands to be sent to fabric components by the external application.
The combination of VIEO Fabric Manager and VIEO FabricView provides OEM
partners with a fully supported, end-to-end InfiniBand management software stack
— from the chipset control software to the management console.
VIEO FabricView includes several modules including Fabric Executive, an
application that aggregates fabric management data, serving as a single source for
requests from external applications, an API Library, that contains a set of
development tools for software developers, VIEO FabricView Console, which
allows access to management facilities using a JAVA-enabled browser, and
translation modules that allow external SNMP messages to pass through the
InfiniBand fabric. As with the Fabric Manager, third-party applications may also be
developed by VIEO partners and can use the VIEO FabricView API Library to
access the VIEO Fabric Executive in the same way as VIEO developed applications.
Funding History
On February 5, 2001, VIEO raised $20 million in Series A funding from investors
including TL Ventures, One Liberty Ventures, Audax Ventures Fund, and Eyes of
Texas Partners. The $20 million in capital will be used to grow VIEO’s software
development and support teams as well as marketing and sales efforts to secure
additional partnerships, and to expand the company’s infrastructure at its Austin and
Richardson, Texas, and St. Paul, Minnesota offices. Moreover, VIEO just recently
announced an additional $4 million in Series A funding from BMC Software and
Dell.
Bringing It All Together
VIEO is conducting pre-beta tests of its technology, with plans to deploy a full beta
test with a number of undisclosed companies later this year. This includes the VIEO
362
The SAN Book III – October 22, 2001
Ethernet
FibreChannel
iSCSI
Host Channel Fabric
NetStorage
Blades
Provide:
Scalable
Bandwidth
and I/O’S
FCAL Drive Interface
Yotta
Storage
Capacity
363
The SAN Book III – October 22, 2001
364
The SAN Book III – October 22, 2001
Zantaz’s Digital Safe has been in development since 1996 and was first introduced
in March, 1999. It is a trusted service for the real-time archiving of high volume
Internet-based email, documents, and transactions. It fully meets the requirements
of the SEC and IRS for companies having regulatory requirements for archiving.
Key benefits include:
➤ outsourced service, no special hardware or software required
➤ highly scalable, allows you to pay only for what is used
➤ instantaneous retrieval with Web-based search engine
➤ transparent to users, no behavioral change required
➤ highly redundant for complete enterprise safety
Digital Safe’s technology draws heavily from the experiences of its founders, who
previously owned and ran traditional paper archiving warehouses. It comprises
highly redundant, geographically diverse data centers that provide complete
enterprise safety. Its proprietary, patent-pending technology brings certified
archiving to companies with regulatory and/or best-business-practice archiving to
their eBusiness operations.
Zantaz has current business relationships with Ernst & Young, Hewlett-Packard,
Cisco (as a member of the Hosted Applications Initiative), Sun Microsystems (as an
“executive” member of Sun Microsystems’ “service provider” program), Critical
Path, and Pilot Network Services, Inc.
Zantaz targets any company for which the retention and retrieval of documents is
part of their long-term business and legal programs and practices. This includes
companies such as securities dealers, brokers, and investment managers, and
applications such as online consumer financial products, corporate e-mail, storage
and data management.
Zantaz has a strategic alliance with Critical Path, a provider of Internet messaging
infrastructure, to provide an Internet-based messaging and eBusiness Archiving
service to the financial industry and corporate markets. The strategic alliance
expands distribution of the Zantaz Digital Safe services to Critical Path’s new and
existing corporate customers who will benefit from long-term email archiving
services.
365
The SAN Book III – October 22, 2001
366
The SAN Book III – October 22, 2001
Glossary
367
The SAN Book III – October 22, 2001
Glossary
100Base-T Fast Ethernet A 100 Mbps technology based on the Ethernet/CD network access method.
10BaseT The IEEE 802.3 specification for ethernet over unshielded twisted pair (UTP).
Agent In the client-server model, the part of the system that performs information preparation
and exchange on behalf of a client or server application. In SNMP, the word agent refers
to the managed system.
Alarm vs. Alert In the most precise usage, an alarm is an off-normal condition (or state) indicating a
performance or equipment degradation or failure. An alert is an indication of a change
(or status) from a normal to an alarm condition or vice versa. However, the term “alarm”
is often used where “alert” would be more accurate. For example, “alarm dialout”
actually occurs on changes from alarm to normal as well as from normal to alarm.
ANSI American National Standards Institute. Standards-setting body responsible for the Fibre
Channel standard.
Application Layer Layer 7 of the OSI Reference Model; implemented by various network applications
including file transfer, electronic mail, and terminal emulation.
Application Program Interface (API) Means of communication between programs to give one program transparent access to
another.
Arbitrated Loop (FC_AL) A Fibre Channel topology that provides a low-cost solution to attach multiple
communicating devices in a single loop. It can support up to 126 nodes and one switch-
attached node in a single loop.
Archive To erase files from primary storage (disk), after a copy has been stored on tape or other
secondary media. The intent is long term storage where the file contents remain
unchanged. This function is also referred to as “retirement” or “shelving.”
Areal Density Bits per square inch of disk surface. Obtained by multiplying the bit density (bits per
inch) by track density (tracks per inch).
Array Management Software The body of software that provides common control and management for a disk array.
Array Management Software most often executes in a disk controller or intelligent host
bus adapter, but may also execute in a host computer.
Asynchronous I/O Requests I/O requests which bear no relationship to each other in time. Asynchronous I/O
requests may overlap in time. Independent access RAID arrays are able to provide
concurrent execution of asynchronous I/O requests which overlap in time, and whose
specified data addresses map to different member disks.
Asynchronous Operations Operations which bear no relationship to each other in time. Asynchronous operations
may overlap in time. The concept of asynchronous I/) operations is central to the use of
independent access arrays in throughput-intensive applications.
368
The SAN Book III – October 22, 2001
Asynchronous Transfer Mode (ATM) (1) The CCITT standard for cell relay wherein information for multiple types of services
(voice, video, data) is conveyed in small, fixed-size cells. ATM is a connection oriented
technology used in both LAN and WAN environments. (2) A fast-packet switching
technology allowing free allocation of capacity to each channel. The-SONET
synchronous payload envelope is a variation of ATM. (3) ATM is an international ISDN
high speed, high-volume, packet switching transmission protocol standard. ATM
currently accommodates transmission speeds from 64 Kbps to 622 Mbps. (4) Permits
hundreds of megabits per second to move on-demand throughout LANs, WANs and
desktops. Uses fixed length cells. Highly scalable, making it suitable for many different
kinds of traffic, from data and voice to video.
Autoloader A device that automatically inserts cartridges into a tape drive. Automation provides
unattended operation and helps ensure that back-ups take place on schedule.
Availability The probability that a system is available at time. Averaged over time, this function
expresses the expected percentage of time a system is available to do useful work.
Backbone (1) The part of a network used as the primary path for transporting traffic between
network segments. (2) A high-speed line or series of connections that forms a major
pathway within a network.
Backup (1) A copy of computer data that is used to recreate data that has been lost, mislaid,
corrupted, or erased. (2) A duplicate copy of data placed in a separate, safe “place”--
electronic storage, on a tape, on a disk in a vault--to guard against total loss in the
event the original data somehow becomes inaccessible. Generally for short-term safety.
Backup File Consolidation The ability for the backup software to recreate a full backup every time an incremental
backup is performed.
Backup Server A computer and storage system which provides backup and restore services for the
network as a network resource.
Backup Version A file, directory, or file space that a user has backed-up, which resides in an backup
storage pool in data storage. Though there may be more than one backup version of a
file in the storage pool, only one is considered the active version.
Backup-Archive Client A program that runs on a workstation and provides a means for users to backup,
archive, restore, and retrieve files. Contrast with Administrative Client.
Bandwidth (1) Measure of the information capacity of a transmission channel. (2) The difference
between the highest and lowest frequencies of a band that can be passed by a
transmission medium without undue distortion, such as the AM band--535 to 1705
kilohertz. (3) The maximum amount of data that can be transmitted through a data
channel per unit time. Usually expressed in megabytes per second. Analog bandwidth
is the range of signal frequencies that can be transmitted by a communication channel
or network. (4) A term used to indicate the amount of transmission or processing
capacity possessed by a system or a specific location in a system (usually a network
system).
Bridge/Router A device that can provide the functions of a bridge, router, or both concurrently.
Bridge/router can route one or more protocols, such as TCP/IP and/or XNS, and bridge
all other traffic.
369
The SAN Book III – October 22, 2001
Browser (1) Term used to describe the client program for the World-Wide Web. Popular browsers
include Mosaic and Netscape. Sometime called “navigator.” (2) Software program that
retrieves, displays, and prints information and HTML documents from the World Wide
Web.
Buffer A portion of storage used to hold input or output data temporarily.
Cache Memory A portion of memory dedicated to collecting and holding related data until a processing
or a storage module is ready to process it. The intent is to improve the overall system
performance. Cache for a disk drive is usually implemented as fast semiconductor
memory. See Read Cache and Write Cache.
Caching File System This is a process that operates on top of the file system to improve performance, data
accessibility, and client storage capacity. Client files are continuously copied to a
designated server so that they are backed-up and always redundant. Inactive files are
removed from the client to improve performance. Client desktop and portable file
systems are maintained and kept synchronized with the server.
CD/Compact Disc An optical disc that stores up to 650MB of data or 74 minutes of audio. CDs can hold
text, photo, image, audio and video files.
CD-R/Compact Disc Recordable Recordable CD technology that allows a user to write to a disc one time.
CD-RW/Compact Disc ReWritable Rewritable CD technology that allows a user to write to the same disc many times. HP
CD-Re-Writable media can be written up to 1,000 times per sector per disc.
370
The SAN Book III – October 22, 2001
Client/Server (1) The relationship between machines in a communications network. The client is the
requesting machine, the server the supplying machine. Also used to describe the
information management relationship between software components in a processing
system. (2) A system architecture in which one or more programs (clients) request
computing or data services from another program (server).
Compression Reducing the size of a data set to lower the bandwidth or space required for
transmission or storage.
Concurrency The ability of backup software to allow a storage device to receive data from more than
one system at a time; interleaving the data streams.
Controller A physical module that interprets signals sent between the host processor and a
peripheral device.
DASD Direct Access Storage Device. A term used by IBM in reference to disk drives.
DAT/Digital Audio Tape A 4mm helical scan magnetic tape technology originally designed for music and later
adapted for backing up computer-created data. DAT drives use cartridges that conform
to the DDS standard for data storage.
Data Accessibility A term defining the system requirements for network storage as continuous access to
files and data, high performance and protection from data loss.
Data Availability An application’s ability to access correct data in a timely manner. Usually implies
continued data availability in the presence of a component failure.
Data Compression Reducing the amount of electronic “space” data takes up. Methods include replacing
blank spaces with a character count, or replacing redundant data with shorter stand-in
“codes.” No matter how data is compressed, it must be decompressed before it can be
used.
Data Reliability Expressed in Mean Time to Data Loss (MTDL), is the average length of time over which
data stored in a disk subsystem can be correctly retrieved.
Data Transfer Rate The rate at which data is transferred to or from the storage media. It is usually given in
thousands of bytes per second (kbytes/second).
371
The SAN Book III – October 22, 2001
Database Backup The software has the capability to backup open relational databases. This is achieved by
a continuous monitor and backup of disk writes (raw partition) the backup of log files
which are files containing the continuous changes to the open database, or API
(Application Program Interface) database backup that allows the use of the native
backup process for the database application. Both methods are analogous to an
incremental backup policy that calls for an initial full backup only once and incremental
backups thereafter at intervals set by the system administrator.
Database Structure The underlying database engine on which the file and media database runs.
Density Generally, recording density.
DDS/Digital Data Storage The most widely used storage standard. DDS is the computer specification for 4mm
DAT technology. This standard was codeveloped by Hewlett-Packard and Sony.
Direct Access Storage Device Any on-line data storage device. A disc, drive or CD-ROM player that can be addressed
(DASD) is a DASD.
Director (Fibre Channel Director) A superior class of switching device. The term is not widely used in the Open System
world, and was introduced by McData as an extension of the IBM ESCON Director.
Disaster Protection A set of rules and procedures that allow a computer site to be put back in operation
after a disaster has occurred. The concept of moving backups off-site constitute the
minimum basic precaution for disaster protection. The remote copy is used to recover
data when the local storage is unaccessible after the disaster.
Disk A non-volatile, randomly addressable, rewritable mass storage device. This definition
includes both rotating magnetic and optical disks and solid-stale disks, or non-volatile
electronic storage elements. It does not include specialized devices such as write-once-
read-many (WORM) optical disks, nor does it include so called RAM disks implemented
using software to control a dedicated portion of a host computer’s volatile random
access memory.
Disk Array A collection of disks from one or more commonly accessible disk subsystems,
combined with a body of Array Management Software. Array Management Software
controls the disks and presents them to the array’s operating environment as one or
more virtual disks.
Disk Mirroring A fault-tolerant technique that writes data simultaneously to two hard disks using the
same hard disk controller. The disks operate in tandem, constantly storing and updating
the same files. Mirroring alone does not ensure data protection. If both hard disks fail at
the same time, you will lose data.
Disk Striping A type of disk array mapping in which consecutive stripes of data are mapped round-
robin to consecutive array members. A striped array, also known as a RAID Level 0
array, provides high I/O performance at low cost, but provides lower net data reliability
than any of its member disks.
Disk Subsystem A collection of disks and the hardware required to connect them to one or more host
computers. The hardware may include an intelligent controller or the disks may attach
directly to a host computer’s I/O bus adapter.
372
The SAN Book III – October 22, 2001
Disk-Based Backup The process of sending backup data to hard disk storage rather than directly to tape.
The backup storage pool on disk is managed for fast restore capability. Background
copies are made to tape or to a remote vault.
Distributed Computing Environment An architecture of standard programming interfaces, conventions, and server function
(DCE) personalities (e.g., naming, distributed file system, remote procedure call) for
distributing applications transparently across networks of heterogeneous computers.
Promoted and controlled by the Open Software Foundation (OSF), a vendor consortium.
DLT/Digital Linear Tape Quantum’s linear magnetic tape technology, used in many high-performance, high-
capacity data backup solutions.
Duplexing A configuration in which each element of a system or subsystem is duplicated, in
addition to the disk drive, example a controller, a host adapter, a power supply. See
Mirroring.
DVD/Digital Versatile Disk (or A CD-sized optical disc with potentially up to 14 times the capacity of a CD (when both
Digital Video Disc) sides of the DVD are used). A single-sided DVD disc might store an entire movie or a full
set of reference books.
EIDE/Enhanced Integrated Device An extension to an IDE interface that supports different IDE device types, such as hard
(Drive) Electronics disk drives, CD-ROM drives and tape drives.
Encryption Data can be coded into an unreadable form and decoding is restricted by password,
user access, or other means. Some software allows decoding to be transparent to the
authorized user, while to the unauthorized user, data is unreadable.
End User Any customer of an interstate or foreign telecommunications service that is not a
carrier, except that a carrier other than a telephone company shall be deemed to be an
“end user” when such carrier uses a telecommunications service for administrative
purposes. It is also a person or entity that offers telecommunications services
exclusively as a reseller shall be deemed to be an “end user” if all resale transmission
offered by such reseller originate on the premises of such reseller.
Enterprise Data Management This is a class of software that builds on Network Data Management software, adding
Software the capabilities of interoperating and providing backup and restore services across
multiple, heterogeneous computing environments. Backup and restore is performed
between any combination of PC’s, UNIX, Midrange systems, and mainframes.
Enterprise Network A geographically dispersed network under the auspices of one organization.
Enterprise Storage Management A family or suite of related processes and applications which manage storage devices
as enterprise resources across heterogeneous environments. Storage management
applications range from asset management to hierarchical storage management.
Ethernet A 10 Mbps, 100 Mbps, and 1 Gbps baseband local area network protocol that allows
multiple stations to access the transmission medium at will without prior coordination. It
resolves contention by using carrier sense and multiple access with collision detection
(CSMAICD). Defined under the IEEE 802.3 protocol.
Exchange A group of sequences which share a unique identifier is called an exchange. All
sequences within a given exchange use the same protocol
373
The SAN Book III – October 22, 2001
F_Port Fabric port. A port on a fabric switch to which N_Ports may be directly connected. An
F_Port is uses the address identifier hex’FFFFFE’.
Fabric A Fibre Channel network consisting of multiple devices interconnected by one or more
switches that use Fibre Channel methodology to link nodes and route frames.
Fabric Switch In this category of switch, any port on any switch can provide (subject to bandwidth
availability) full speed access to any other port on the network. The network consists of
a fabric of linked switches.
Failure A detectable physical change to hardware. Failures may be repaired by the replacement
of a physical component.
Fast LAN Term given to two emerging standards; IEEE 802.3u (called Fast Ethernet) for Ethernet
operating at 100 Mbps over Cat-3 or 5 UTP, and IEEE 802.12 (called 100VG-AnyLAN) for
Ethernet or Token Ring operating at 100 Mbps over CAT-3/4/5 UTP, STP or optic fiber.
Fault tolerance The ability for a system to remain operational, sometimes in a degraded mode, even
though one or more elements of the system have failed. Fault tolerance is obtained
through the use of redundant components or modules.
FDDI II The proposed ANSI standard to enhance FDDI. FDDI II will provide isochronous
transmission for connectionless data circuits and connection-oriented voice and video
circuits.
Fiber Distributed Data Interface An emerging high-speed networking standard. The underlying medium is fiber optics,
(FDDI) and the topology is a dual-attached, counter-rotating Token Ring. FDDI networks can
often be spotted by the orange fiber “cable.” The FDDI protocol has also been adapted
to run over traditional copper wires.
Fibre Channel Arbitrated Loop This is one of the possible physical topologies of Fibre Channel. In this topology, the
(FC/AL) Fibre Channel is connected in a loop with devices all connecting to the loop. It can be
thought of as a similar structure to a token ring network. Up to 126 nodes can be
connected to the loop.
Fibre Channel Fabric This is a structure where addressing of ports on a network of Fibre Channel is made
independently of the physical location or address of the target port. Switches are
responsible for passing Fibre Channel packets to the target port regardless of which
Fibre Channel loop or switch the port physically resides on. Multiple switches can be
connected to create large networks with up to 224 addressable ports.
Fibre Channel Point-to-Point This topology for Fibre Channel provides a simple direct connection between just two
nodes. This is the closest approximation to the traditional SCSI topology.
Fibre Channel Topology A number of possible topologies exist for the connection of Fibre Channel. One is point-
to-point, where a direction connection is made between nodes on the Fibre Channel.
Another is Arbitrated Loop where multiple nodes can be connected together. Lastly
there is a Fibre Channel Fabric topology which allows for multiple loops and point-to-
point connections to be combined into networks using switches.
374
The SAN Book III – October 22, 2001
File Allocation Table (FAT) Data written to a magnetic disk is not necessarily placed in contiguous tracks. It is
usually divided into many clusters of data in many locations on the disk surface. The
FAT is the special area on a disk which keeps track of where clusters of data have been
written for later retrieval later.
File and Software Distribution There are many levels to this application set. The simplest is the ability to recognize
revision levels of software installed on the network and to broadcast upgrades so that
revision levels can be maintained automatically.
File Migration File migration is an automated process where candidate files are removed from primary
storage, leaving a phantom (or stub) file in place of each file. Candidate files are
automatically identified for removal based on their inactivity or file type. When the
phantom file is read, it triggers a transparent retrieval from the secondary storage pool
File Retrieval The return of migrated files from the lower level of the storage hierarchy to active
storage in response to user access request.
File Server Local Area Network (LANs) were invented to allow users on the LAN to share and
thereby conserve the cost of peripherals (printers, modems, scanners) and to likewise
share software. The file server is the machine on the LAN where the shared software is
stored.
File System Part of an OS that tracks, retains and monitors all files. The FS can be assessed by other
parts of the OS such as the computer back plain and storage along with applications via
well-documented APIs.
File Tracking The ability of the on-line catalogue to provide complete file versioning and history
information along with specific media locations.
File Transfer Protocol (FTP) (1) An IP application protocol for transferring files between network nodes. (2) An
Internet protocol that allows a user on one host to transfer files to and from another host
over a network.
File Transfer, Access, and The OSI remote file service ad protocol.
Management (FTAM)
File-Oriented Backup Any backup software that instructs the computer to store information in files just as they
appear on the originating computer, making restoration easier and more logical.
FINL_Port An NL_Port which is capable of providing certain Fabric services to other NL_Ports on a
Loop in the absence of a Fabric. This NL_Port will respond to requests to open
communication with AL_PA hex’00’, even though it may actually have another value for
its AL_PA.
FL_Port Fabric-Loop port. An F_Port which is capable of supporting an attached Arbitrated Loop.
An FL_Port on a Loop will have the AL_PA hex’00’, giving the Fabric highest priority
access to the Loop. An FL_Port is the gateway to the Fabric for NL_Ports on a Loop.
Flow Control Flow control is a method of exchanging parameters between two connected devices and
managing the rate of frame transmission. Fibre Channel uses credit-based flow control
to prevent the possibility of frames being transmitted so rapidly that the buffers overflow
and lose data. Flow control. Enables concurrent multiple exchanges. Some Classes of
Service use end-to-end flow control to ensure that lost frames are detected.
375
The SAN Book III – October 22, 2001
Frame Relay High-performance interface for packet-switching networks. Considered more efficient
than X.25 which it is expected to replace. Frame relay technology can handle “bursty”
communications that have rapidly changing bandwidth requirements.
Frame Switch A device, similar to a bridge, which forwards frames based on the frame’s layer 2
address. Frame switches are generally of two basic forms, cut-through switch (on the
fly switching), or store and forward switch. LAN switches such as Ethernet, Token Ring
and FDDI switches are all examples of frame switches.
Full Backup A backup of all files on a storage device. If files or directories have been excluded from
backups, they will not be contained in a full backup.
Gigabit (Gb) One billion bits.
Gigabyte (GB) (1) A billion (slightly more) bytes of data, or a thousand megabytes. Presently, gigabytes
are in the stratosphere of storage capability, but imaging applications commonly take up
huge amounts of data. For example, it only takes ten 8 1/2” by 11” color pictures,
9
scanned at 600 dpi, to fill a gigabyte. (2) Shorthand for 1,000,000,000 (10 ) bytes. Some
9
prefer to use the (10 ) convention commonly found in I/O-related literature rather than
30
the 1,073,741,824 (2 ) convention sometimes used in describing computer system
memory.
Graphical User Interface (GUI) A type of user interface that includes a combination of graphics and the use of pointing
devices, menu bars, overlapping windows and icons.
Hardware The mechanical, magnetic and electronic components of a system, e.g., computers,
telephone switches, terminals and the like.
HBA See Host bus adapter.
Hierarchical File System (HFS) In DOS, the file management system that allows directories to have subdirectories, and
sub-subdirectories. In Macintoshes files may be placed into folders, and folders to be
placed within other folders.
376
The SAN Book III – October 22, 2001
Hierarchical Storage Management The process of automatically storing data on the lowest cost devices that can support
(HSM) the performance required by the applications. To users, data storage never fills and file
access, regardless of location in the storage hierarchy, is completely transparent. The
software automatically manages multiple levels of storage hierarchy primary, secondary
(nearline), or off-line
Host Backup Server The backup server that is the central repository (vault) in an enterprise data
management architecture. Local backup servers ship data to the host backup server.
Centralized administration is performed and controlled from the host backup server. The
host backup server can be located anywhere in the enterprise and on any platform in
which will support the requirements.
Host Bus Adapter A host bus adapter allows to connect a server system bus with an external device.
Host Computer Any computer system to which disks are directly attached and accessible. Mainframes,
and servers, as well as workstations and personal computers, can all be considered
host computers in the context of this book, as long as they have disks attached to them
Host-Based Array A disk array whose Array Management Software executes in its host computer(s) rather
than in a disk subsystem.
Hot Backup A backup performed on a local disk drive as opposed to (or in addition to) the classical
backup to tape. It allows faster and more frequent data transmission, and quicker
recovery.
Hot Spare Or On-line spare. A drive that is present in the system but normally unused until another
drive fails, at which time the hot spare is automatically substituted for the failed drive.
Hot Swap The substitution of a replacement unit (RU) in a disk sub-system for a defective one,
where the substitution can be performed while the subsystem is running (performing its
normal function). Hot swaps are manual performed by humans
Hub This is a simple connectivity device that allows for devices to be connected to a fibre
channel loop by being attached to a hub port. The advantage of this is that failures of a
single device on the loop can be isolated from the other ports on the loop. The
aggregate bandwidth of the hub is still that of a single fibre channel loop however.
IDE/Integrated Device (Drive) A common hard drive interface built into most 386/486 PCs. This interface has been
Electronics replaced by EIDE in Pentium and newer PCs.
I/O - Input/output Input/Output. Term applied to a channel or device ( such as an EIDE, parallel port or
SCSI connection) that can perform an input process and an output process at the same
time.
I/O Driver A host computer software component (usually part of the operating system) whose
function is to control the operation of peripheral controllers or adapters attached to the
host computer. I/O drivers communicate between applications and I/O devices, and in
some cases may participate in data transfer, although this is rare with disk drivers,
since most disk adapters and controllers contain hardware for data transfer.
377
The SAN Book III – October 22, 2001
I/O Load The sequence of I/O requests made to one or more disk subsystems by a host
computing environment. The host computing environment includes both applications
and host overhead functions, such as swapping, paging, and file system activity.
I/O Load Balancing The adjustment of system and/or application components so that application I/O
demands are spread evenly across the I/O system’s physical resources. I/O load
balancing may be done manually (by a human) or automatically (by some means that
does not require human intervention).
I/O Operations per Second (IOPS) A generic measure of I/O performance. To be meaningful, the type and operation
mixture of I/O must be specified as well, such as read IOPS.
I/O-Intensive A characterization of applications. An I/O-intensive application is one whose
performance depends strongly on the performance of the resources that provide its I/O
services.
Incremental Backup An operation that backs up all data that has been modified or added since a given date.
The date is usually the date of the last full or incremental backup.
Input/Output (I/O) Refers to the process, techniques and media used for human/machine communication.
Also refers to data submitted between computing components.
Intermix A mode of service defined by Fibre Channel that reserves the full Fibre Channel
bandwidth for a dedicated Class 1 connection but also allows connectionless Class 2
and 3 traffic to share the link if the bandwidth is available.
IP Number/IP Address Also referred to as a “dotted quad.” A code number consisting of 4 parts separated by
dots, e.g., 165.113.245.2. Every computer on the Internet has its own unique IP
number.
IPI Intelligent peripheral interface.
Isochronous Transmission Data transmission which supports network-wide timing requirements. A typical
application for isochronous transmission is a broadcast environment which needs
information to be delivered at a predictable time.
JBOD Just a Bunch of Disks. Refer to a disk drive configuration in which there is no
redundancy.
Jukebox A device that holds multiple optical disks and one or more disk drives, and can swap
disks in and out of the drive as needed. Same as an autochanger.
L_Port Loop port. Generic term for an NL_Port or FL_Port i.e., any Fibre Channel port which
supports the Arbitrated Loop topology.
Large I/O Request An I/O, read, or write request that specifies the transfer of a large amount of data.
Large Read Request ‘Large’ usually depends on the context, but often refers to 64 KBytes or more.
Large Write Request
378
The SAN Book III – October 22, 2001
Latency A measurement of the time it takes to send a frame between two locations. Low latency
is a fundamental requirement for storage applications and is typical of I/O channel
technologies. Fibre Channel connections are characterized by low latency. In a disk
drive, latency relates to the disks rotational speed and must be considered in
determining a disk drives total access time.
Library Manager A library manager sits as the piece of middleware between a backup, archive, or some
other application and an automated media repository such as an optical jukebox or tape
library. The library manager intercepts the call for a specific file, looks up the files’
location (which piece of media, surface block ID) and issues commands to the jukebox
or library manager to mount a specific media and retrieve the data requested. The
library manager contains the file history databases, media location database, device
drivers for the supported media repositories and other management functions. Media
management capabilities are usually built into a library manager. It is a superset of
functions to a media manager.
Life Cycle Management A process of tracking and monitoring the number of times that the media is recorded or
certain zones are overwritten (this translates to the number of passes over the heads).
Management also requires the monitoring of raw error rates in order to prevent
excessive usage and data loss.
Local Area Network (LAN) (1) A network covering a relatively small geographic area (usually not larger than a floor
or small building). Compared to WANs, LANs are usually characterized by relatively high
data rates. (2) Network permitting transmission and communication between hardware
devices, usually in one building or complex. (3) High-speed transmissions over twisted
pair, coax, or fiber optic cables that connect terminals, personal computers, mainframe
computers, and peripherals together at distances of about one mile or less.
Local Backup Server A CPU and storage pool which acts as a centralized backup device and repository for a
set of workstations and servers in a network. Multiple local backup servers are tied
together and act synchronously to provide backup for the network under central control.
Data is ported offsite to the host backup server. Administration is performed, centrally,
for all local backup servers either locally or from the host backup server.
Logical Disk A set of contiguously addressed member disk blocks that is part of a single virtual disk-
to-member disk mapping. Logical disks are used in some array implementations as
constituents of logical volumes or partitions
Logical Volume A virtual disk made up of logical disks. Also called a virtual disk, volume set, or partition.
Magneto-Optic (MO) Recording Recording data using optical means to change the polarity of a magnetic field in the
recording medium. Data is erasable and/or rewritable.
Managed Hub This is a technique for providing statistics information about the traffic on a hub. No
actual management of the hub is usually possible using this interface, but information
and notification of failures can be achieved. This interface often uses Simple Network
Management Protocol (SNMP) Management Information Bases (MIBs) as a standard
protocol for providing this information.
379
The SAN Book III – October 22, 2001
Mapping The conversion between multiple data addressing schemes. Most commonly used to
refer to conversion between member disk block addresses and block addresses of the
virtual disks presented to the operating environment by Array Management Software.
Mean Time Between Failures A measure of equipment reliability, the higher the MTBF, the more reliable the
(MTBF) equipment.
Media Plural of medium. The physical environment through which transmission signals pass.
Common network media include twisted pair, coaxial and fiber optic cable, and the
atmosphere (through which microwave, LASER, and infrared transmission occurs).
Media Management The software will monitor the number of write passes or use other techniques to
monitor media quality, aging, and recommend new media and/or suggest using the
media for another purpose such as archiving. It is a pro-active and automated feature
allowing preventive media failures and requiring less system administrator time in
manual media management and error control.
Media Types Managed Optical or tape media are common. There are many types of tape media that are
supported by libraries; 3490, DLT. 8 MM. 4 MM, and VHS (VLDS).
Megabits Per Second (Mbps) A digital transmission speed of millions of bits per second.
380
The SAN Book III – October 22, 2001
Near-Line Data that is available on a secondary storage device that the user can access, but at a
slower rate than the on-line data is accessed.
Nearline Storage Or Near Online Storage. Storage that can be accessed online only through the use of
robotics. It does not require human intervention, and can be accessed much quicker
than off-line storage, though not as quickly as online prime storage.
Network (1) A collection of computers and other devices that are able to communicate with each
other over some network medium. (2) An aggregation of interconnected nodes,
workstations, file servers, and/or peripherals, with its own protocol that supports
interaction.
Network Analyser A hardware/software device offering various network troubleshooting features, including
protocol-specific packet decodes, specific pre-programmed troubleshooting tests,
packet filtering, and packet transmission.
Network Attached Storage (NAS) This is the provision of storage in a form that is readily accessible on a network.
Traditionally this meant having a storage array device that incorporated a file system
and presented either a networked file system protocol such as NFS, or else emulated a
disk device so that the array could be connected to a storage I/O interconnect to the
host.
Network Availability A measure of network uptime. High network availability is a term designed to convey
the message that 100% network uptime is very desirable. Most networks only achieve
96% uptime due to planned and unplanned outages.
Network Backup Software Backup software capable of providing backup and restore of network files for servers
and clients. This software is capable of backup and restore only. It is flexible and has
many features such as tape catalogue s, tape libraries, and many storage devices.
Network Data Management Backup software which includes high performance backup and restore features plus
Software manages data and files in the network. Features such as disk-based database/librarian,
automated archiving, and file migration set it apart from Network Backup.
Network File System (NFS) (1) A protocol defined by Sun Microsystems that that extends TCP/IP
network file services. NFS permits remote node files to appear as though
they are stored on a local workstation. (2) A distributed file system in
UNIX developed by Sun Microsystems which allows a set of computers
to cooperatively access each other’s files in a transparent manner.
Network Storage Management (1) Network Storage Management is not an application. Rather, it is a family of related
processes and application that manage storage devices as network resources. Storage
management applications range from asset management to hierarchical storage
management. (2) A system responsible for managing at least part of a network. NMSs
communicate with agents to help keep track of network statistics and resources.
Network Storage Resource A storage pool that can be accessed by every client in the network. Its access is
completely transparent to the user, whether the storage is online or nearline.
Network Topology Physical arrangement of nodes and interconnecting communications links in networks
based on application requirements and geographical distribution of users.
NL_Port Node-Loop port. An N_Port which can operate on the Arbitrated Loop topology.
381
The SAN Book III – October 22, 2001
NOS Network Operating System; UNIX, WIndows NT, and Novell Netware are examples.
Off-Line Data that is not physically stored on an accessible drive, such as removable tapes or
disks.
On-Line Data that is available on a primary storage device so that it is readily accessible to the
user.
Open File Backup The backup of open files (which may be modified during backup). The file must remain
static for backup but available for reading and writing.
Open Network Architecture (ONA) The structure of a communications network that permits all types of terminals and users
to have access to any network capability without compatibility restrictions.
Operating system Collection of programs that, taken together, manage the hardware and software; it is
the operating system that makes the hardware usable, providing the mechanisms that
application programs use to interact with the computer.
Optical (1) Containing lenses mirrors, etc., as in optical view-finder and optical printer. (2) In
general, having to do with light and its behavior and control, as in optical properties,
optical rotation. (3) Pertaining to the science of light and vision.
Optical Disk A storage device that is written and read by laser light. Certain optical disks are
considered Write Once Read Many (WORM), because data is permanently engraved in
the disk’s surface either by gouging pits (ablation); or by causing the non-image area to
bubble, reflecting light away from the reading head. Erasable optical drives use
technologies such as the magneto-optic technique, which electrically alters the bias of
grains of material after they have been heated by a laser. Compact disks (CDs) and
laser (or video) disks are optical disks. (When referring to CD technology, the spelling
disc is used. In all other cases, the spelling disk is used.)
Optical Drive Machine for reading or writing a data storage medium (disk, tape, card, etc.) that uses
light for examining patterns.
Parallel Access Array A disk array in which the data mapping and protection model assumes that all member
disks operate in unison, and that all member disks participate in the execution of every
application 1/0 request. Parallel access arrays may require physical disk
synchronization, or they may approximate true parallel access behavior.
Parallel/Concurrent Parallel is the capability to manage multiple jobs (backup/restore) at once by keeping
multiple tape drives operating in parallel. Concurrence allows a storage device to
receive data from more than one system at a time, interleaving the data streams.
Partial File Backup A backup in which only the changes to a file (usually a large file) are copied to backup
media instead of copying the entire file. This assumes that an older version of the entire
file has already been copied to backup media.
Partition A virtual disk. The term partition is most often used when a redundancy group is
presented to a host operating environment as more than one virtual disk or volume set.
Also used in complex arrays with dynamic mapping to denote a collection of redundancy
groups dedicated to providing the storage for a subset of an array’s virtual disks.
382
The SAN Book III – October 22, 2001
Physical Media Any physical means for transferring signals between OSI systems. Considered outside
the OSI Model, and sometimes referred to as “Layer O,” or the bottom of the OSI
Reference Model.
Point-to-Point Pertaining to a direct link that supports data transmission between two nodes.
Ports Fibre Channel ports come in a number of flavours depending on the topology of the
Fibre Channel
F_Ports are point-to-point ports connected to a fabric. Generally this means that
the F_Port is a port on a switch.
FL_Ports are ports connecting from one loop to a switch and thus to a fabric
G_Ports are classified by some switch companies as ports that can be either
E_Ports or F_Ports depending on usage.
Protocol Set of data communication conventions, including timing, control, formatting, and data
representation. See Transport protocol
QIC/Quarter Inch Cartridge A magnetic tape technology used for backup. QIC tape comes in 3.5-inch minicartridges
and 5.25-inch data cartridges.
RAID Level A number designating the general configuration of an array. RAID configurations are
defined and generally accepted from level 0 through 5.
Random Access The ability to position data (as in disk) without reading intervening data blocks (as in
tape).
Random I/O An I/O load whose consecutively issued read or write requests do not specify adjacently
Random Reads located data. Random I/O is characteristic of I/O request-intensive applications.
Random Writes
Read Cache A cache segment dedicated to saving information read from the disk drives, on the
assumption that this data will be soon requested again by the system. The system will
initiate further disk accesses only if the desired information is not located in the cache.
Read Only Memory (ROM) Data stored in a medium that allows it to be accessed but not erased or altered.
Read/Write Ratio The ratio of read operations to write operations in a typical host system workload. This
is important in selecting RAID configurations, as some configurations are inefficient in
write intensive environments.
Read-Only Access to Migrated Data The normal retrieval process automatically writes the file back to primary storage. If the
user is scanning a large number of files, this procedure can rapidly overload the primary
file system. An alternative is to read the files from the migration storage pool directly
into memory and not write to disk unless the user chooses to do so. This way, the
unused files can be discarded without impacting disk storage.
383
The SAN Book III – October 22, 2001
Reconstruction The regeneration of all of the data from a failed disk in a RAID Level 1, 3, 4, 5, or 6 array
onto a replacement disk. Reconstruction normally occurs without interruption of
application access to data stored on the array’s virtual disks.
Reduced Mode A mode of RAID array operation in which not all components of the array are
functioning, but the array as a whole is functioning properly (responding to application
read and write requests).
Reduction The removal of a member disk from a RAID array. Reduction most often occurs because
of member disk failure, however, some RAID implementations allow reduction for
system management purposes.
Redundancy The utilization of modules that are not necessary for a normal system operation. They
get substituted to (or can perform the same function as) another component, allowing
the system to remain operational if the component fails.
Redundant Arrays of Inexpensive or A storage device that uses several optical discs working in tandem to increase
Independent Discs (RAID) bandwidth output and to provide redundant backup.
Regeneration The process of rebuilding user data that was stored on a failed member disk of a RAID
Level 1, 3, 4, 5, or 6 array . Regeneration may be used to recover data when a member
disk has failed. It may also be used to recover data when an unrecoverable media error
is encountered on a member disk. Data regeneration in a RAID Level I array consists of
delivering the alternate copy of the data. Data regeneration in a RAID Level 3, 4, or 5
array consists of executing the array’s parity computation algorithm. Data regeneration
in a RAID Level 6 array consists of choosing the more convenient of the two parity
algorithms, rebuilding the data, and adjusting the alternate parity algorithm as required.
Relational Database A database built and operated in accordance with the relational model of data which
holds that all data be organized as a set of two dimensional arrays or tables which have
a relation to each other.
Remote Access The process of allowing remote workers to access a corporate LAN over analog or
digital telephone lines.
Remote Access Server Access equipment at a central site that connects remote users with corporate LAN
resources.
Remote File System (RFS) A distributed file system, similar to NFS, developed by AT&T and distributed with their
UNIX System V operating system.
Remote Vaulting The software has the ability to reach out or transmit over phone or direct connection
with other sites, and manage that data, essentially treating remote site data as a client
system or visa versa.
Remote Vaulting Capability The ability of the media manager to transmit files and data offsite to a remote vault.
Replacement Disk A disk available for use as or used to replace a failed member disk in a RAID array.
384
The SAN Book III – October 22, 2001
Replacement Unit A component or collection of components in a disk subsystem which are always
replaced (swapped—q.v.) as a unit when any part of the collection fails. Abbreviated
RU. A replaceable unit is commonly called a consumer replaceable unit (CRU) if the
replacement can be performed by the end user, and a field replaceable unit (FRU) if the
replacement must be performed by a vendor service representative. Replacement units
may be cold, warm, hot, or auto-swapped (q-v-).
Reporting The ability to develop reports on access, media status, and other management requests.
Request-Intensive A characterization of applications. Also known as I/O -intensive
Restore A function that allows users to copy files from the backup storage pool to an online
storage device.
Retrieve A function that allows users to copy files from the archive storage media to an online
storage device.
Rewritable Optical Optical media from which data can be erased and new data added. Magneto-optical
and phase change are the two main types of rewritable optical disks.
Router (1) An OSI Layer 3 device that can decide which of several paths network traffic will
follow based on some optimality metric. Also called a gateway (although this definition
of gateway is becoming increasingly outdated), routers forward packets from one
network to another based on network-layer information. (2) A dedicated computer
hardware and/or software package which manages the connection between two or
more networks.
Routing Bridge MAC-layer bridge that uses network layer methods to determine a network’s topology.
Scalable Capable of growing through additions of modular increments, without necessitating
major modifications to the original system. In a storage context, a scalable system is
capable of efficiently handling the requirements from a small system to a large system,
in terms of capacity, performance, availability, connectivity, and/or manageability.
SCSI Small Computer Systems Interconnect. A collection of ANSI standards which define I/O
buses primarily intended for connecting mass storage subsystems or devices to hosts
through I/O bus adapters. Originally intended primarily for use with small (desktop and
deskside workstation) computers, SCSI has been extended to serve most computing
needs, and is arguably the most widely implemented I/O interconnect in use today. SCSI
is used for local data communication between a host CPU and an attached SCSI bus that
contains intelligent peripheral devices such as disks, tapes, scanners and printers. As
many as six devices can be connected to a single SCSI channel. Variations include:
Ultra SCSI, a faster SCSI-2 interface that transfers 20MB/sec for narrow 8-bit
versions and 40MB/sec for wider 16-bit versions.
385
The SAN Book III – October 22, 2001
Security Features Security features include operator access privileges, authorization levels, and password
control to limit access to unauthorized files and data.
Seek The physical motion to relocate the arm from one track to another.
Selective Backup A function that allows the backup program to include or exclude certain files from being
systematically backed up during a full or an incremental backup operation.
Sequence A series of related frames, when strung together in numbered order, creates a sequence
which can be transmitted over a Fibre Channel connection as a single operation
Sequential I/O An 110 load consisting of consecutively issued read or write requests to adjacently
Sequential Reads located data. Sequential I/O is characteristic of data transfer intensive applications.
Sequential Writes
Server A computer which is dedicated to one task. A database or directory server would be
responsible for responding to a user’s search request, returning the list of stored
documents that meets with the parameters of the request.
Shelf Storage Media placed external to an automated repository but, nearby so that an operator can
mount it upon request.
Simple Network Management (1) The Internet network management protocol. SNMP provides a means to monitor and
Protocol (SNMP) set network configuration and runtime parameters. (2) A standard protocol that runs
over an IP link to provide management of network type devices without performing
continual polling.
Small Computer System Interface An industry standard for connecting peripheral devices and their controllers to a
(SCSI) microprocessor. SCSI defines both hardware and software standards for communication
between a host computer and a peripheral. Computers and peripheral devices designed
to meet SCSI specifications should work together. A single SCSI adapter card plugged
into an internal IBM PS/2 micro channel PC slot can control as many as seven different
hard disks, optical disks, tape drives and scanners, without siphoning power away from
the computer’s main processor. Formerly known as SASI (Shugart Associates Systems
Interface).
Small I/O Request An I/O, read, or write request that specifies the transfer of a relatively small amount of
Small Read Request data. ‘Small’ usually depends on the context, but most often refers to 8 KBytes or fewer.
Small Write Request
Solid State Disk A disk whose storage capability is provided by solid state random access memory rather
than magnetic or optical media. An electronic storage element generally offers very high
performance, because seek and rotational latency times are negligible, and may offer
very high data transfer rate compared to a conventional disk. Cost per byte of storage,
however, is quite high compared to conventional disks, and volumetric density is lower.
These two factors combine to make solid state disk generally lower capacity than
magnetic or optical disks.
Spare, Spare Disk Informally, a disk or P_extent reserved for the purpose of substituting for a like entity in
case of failure of that entity.
SSA (Serial Storage Architecture) A high speed serial interface designed and marketed by IBM.
386
The SAN Book III – October 22, 2001
Storage Area Network (SAN) Simply put, this is the evolutionary step of connectivity provided by a support (e.g., Fibre
Channel) between host and storage. This connectivity provides a true network of storage
devices and the hosts that access the storage.
Storage Capacity Amount of data that can be contained in an information holding device or main memory,
generally expressed in terms of bytes, characters or words.
Storage Density Usually refers to recording density (bpi, tpi, or a combination of both).
Storage Hierarchy A logical ordering of the backup and archive storage devices, as defined by an
administrator with system privilege. Generally, the ordering is based on the speed and
capacity of the devices. Typically, each storage pool is associated with a different type
of storage device; online-disk, nearline-optical, jukebox tape library or offline-tape.
Storage Media The physical device itself, onto which data is recorded. Magnetic tape, optical disks,
floppy disks are all storage media.
Storage Pool A named collection of storage volumes whether online, nearline, or off line.
Storage Subsystem A disk subsystem (q.v.) whose intelligent controller or I/O bus adapter is capable of
attaching other forms of mass storage such as CDROMs or tape transports.
Striped Array The collection of disks comprising an array which implements the RAID Level 0 or disk
striping (q.v.) mapping technique.
Striping Short for disk striping (q.v.); also known as RAID Level 0. The process of recording data
across multiple disks rather than on one drive. Data is divided into segments each of
which is written to successive drives. Striping is also referred to as RAID 0.
Stub File The record of the original file left on the primary drive after the file itself has been
migrated to a secondary drive. Also called a key or a place-holder. Click on it and the
file appears on your screen as it normally would, but delivered to you at the slower
access speed of the secondary device. The stub file takes up much less space than the
original, unmigrated file. When the user attempts to open a stub file, it copies back the
migrated material to the primary drive, then opens it.
Subsystem-Based Array A disk array whose Array Management Software executes in the controller or host I/O
bus adapter of an intelligent disk subsystem.
Swap The installation of a replacement unit in place of a defective unit. Units are parts of a
disk subsystem which may either field replaceable (FRUs) by a vendor service
representative or consumer replaceable (CRUs). The swap operation itself may either be
a cold, warm, hot, or auto-swap, depending on the state in which the disk subsystem
must be in order to perform it.
Switch — Switched Fabric A Fibre Channel switch routes the data . It has the ability to shift circuit connections as
required to support data transmission requests. When this type of connection is
expanded to include more than one switch and multiple nodes, it is called a switched
fabric
Switched Ethernet Configuration supporting an Ethernet hub with integrated MAC layer bridging or
switching capability to provide each port with 10 or 100 Mbps of bandwidth. Separate
transmissions can occur simultaneously on each port of the switching hub, and the
switch filters traffic based on the destination MAC address.
387
The SAN Book III – October 22, 2001
System Disk The disk on which a system’s operating software is stored. The system disk is usually
the disk from which the operating system is bootstrapped (initially loaded into memory).
The system disk frequently contain the system’s swap and/or page files. It may also
contain libraries of common software shared among several applications.
Systems Network Architecture IBM’s very successful means of networking remotely located computers. It is a tree-
(SNA) structured architecture, with a mainframe host computer acting as the network control
center. Unlike the telephone network, which establish a physical path for each
conversation, SNA establishes a logical path between network nodes, and it routes each
message with addressing information contained in the protocol.
Tape Backup Making mag tape copies of hard disk and optical disc files, for disaster recovery.
Tape Drive The machine, actually a collection of devices, that transports, reads and or writes a
magnetic tape.
TCP/IP Transmission Control Protocol/Internet Protocol. A set of communication protocols that
support peer-to-peer connectivity functions for both local and wide area networks. Also
used for Internet access from the desktop.
Terabyte (TB) From “tera,” which means trillion, although it actually means 1,099,511,627,776 bytes
in a computer’s binary system. A terabyte is 1,024 gigabytes.
Throughput (1) The amount of time it takes for the processing of data from the beginning of a
particular process, to the end of the process. Throughput also can refer to the number
of items completed in the process. (2) The number of I/O requests satisfied per unit
time. Expressed in requests/second, where a request is an application request to a
storage subsystem to perform a read or write operation.
Throughput-Intensive A characterization of applications. Also known as request-intensive. A throughput-
intensive application is an I/O intensive (q.v.) application which makes large numbers of
requests asynchronously, as opposed to a data transfer-intensive (q.v.) application.
Transmission Control (1) A set of protocols developed by the Department of Defense to link dissimilar
Protocol/Internet Program (TCP/IP) computers across networks. (2) A reliable, full duplex, connection-oriented end to end
transport protocol running on top of IP.
Transparent Retrieval Transparent retrieval is achieved by several processes. The objective is for the user to
access all their files and data as if they were on-line. Some methods leave a stub or
phantom file which have the same name as the migrated file. When the application
begins to read the file, the stub triggers a retrieval. Control is returned to the user along
with a message so that they can do something else while the retrieval occurs.
Travan A commonly used tape backup technology that grew out of the QIC standard. Because
of their simplicity and affordability, Travan products tend to be ideal for backup desktop
and portable PCs and smaller servers.
388
The SAN Book III – October 22, 2001
Transport Protocols A set of rules and procedures that support connectivity functions for local or wide area
networks. Examples are:
TCP/IP (UNIX)
IPX/SPX (NetWare)
SNA (IBM)
UDF/Universal Disk Format A file system for CD optical media. UDF was designed for read-write interoperability
between all the major operating systems as well as compatibility between CD rewritable
and CD write-once media.
Unformatted Capacity Storage capacity of disk drive prior to formatting; also called the gross capacity.
Value Added Reseller / Companies that buy equipment from computer or telephone manufacturers, add some
Value Added Dealer of their own software and possibly some peripheral hardware to it, and then resell the
(VAR/VAD) whole computer or telephone system to end users.
Virtual Disk An alternative name for volume set (q.v.). In order to be consistent with ANSI X3T9.2
RAID Study Group terminology, the RAID Advisory Board is gradually replacing this
commonly used term with the term volume set in its publications.
Volume Label A name assigned to a floppy or hard disk in MS-DOS. The name can be up to 11
characters in length. You assign a label when you format a disk or, at a later time, using
the LABEL command.
Volume Management The ability to make two distinct disk drives or tape subsystems appear to the OS as one
logical system. Therefore, if you have 4 disks drives, you can structure those four disks
drives as one large disk farm enabling data to be written to the entire disk farm at the
same time.
Volume Set (An ANSI X3.T10 term) A contiguous range of logical block addresses presented to the
operating environment by Array Management Software. The volume set, sometimes
called a virtual disk (q.v.), is the array entity which most closely resembles a disk when
viewed by the operating environment.
Web Browser Client software that requests and displays HTML documents and other Internet or
intranet resources.
Wide Area Network (1) A network which encompasses interconnectivity between devices over a wide
(WAN) geographic area. Such networks would require public rights-of-way and operate over
long distances. (2) A network linking computers, terminals and other equipment that
covers an area larger than a single building or campus.
389
The SAN Book III – October 22, 2001
Workstation A desktop computer that performs local processing and accesses LAN or WAN services.
Write Cache A cache segment used to accumulate data before writing to the drive on the theory that
a single large write operation is more efficient than several smaller transfers.
Write Once Read Many Optical storage device on which data is permanently recorded. Data can be erased, but
(WORM) not altered, and no additional data can be added.
Write Penalty The low apparent performance of application write requests to independent access RAID
arrays’ virtual disks. The write penalty is inherent to independent access RAID data
protection schemes, which require multiple member writes for each application write
and ranges from minimal (RAID Level 1) to substantial (RAID Levels 5 and 6). Many array
designs include features such as write-back cache specifically to minimize the write
penalty.
Zoning This is the term used by some switch companies to denote the division of a SAN into
sub nets that provide different levels of connectivity or addressability between specific
hosts and devices on the network. In effect routing tables are used to control access of
hosts to devices. This zoning can be performed by cooperative consent of the hosts or
can be enforced at the switch level. In the former case, hosts are responsible for
communicating with the switch to determine if they have the right to access a device.
390
The SAN Book III – October 22, 2001
Table of Figures
391
The SAN Book III – October 22, 2001
392
The SAN Book III – October 22, 2001
393
The SAN Book III – October 22, 2001
394
The SAN Book III – October 22, 2001
395
The SAN Book III – October 22, 2001
396
The SAN Book III – October 22, 2001
Notes
397
The SAN Book III – October 22, 2001
Notes
398
The SAN Book III – October 22, 2001
399
The SAN Book III – October 22, 2001
400