Sei sulla pagina 1di 71

Storage Basics

Agenda/learning objectives
Introduce the components of the computer and show

how they request and store data

Introduce RAID technology and RAID protection

schema

Introduce Storage Area Networks and Network

Attached Storage

Introduce different data protection capabilities available


Show how all the components fit into the Information

Lifecycle Management vision

The Input / Output Machine

OUTPUT

INPUT

Save changes
to Forecast.doc
OUTPUT

INPUT

Open
Forecast.doc
OUTPUT

INPUT

Print this
E-mail

The CPU or Central Processing Unit


(Server)

Where Data Is Stored


Main Memory
Very fast accessno moving parts
Very expensive compared to mechanical or magnetic storage
Volatilerepresents ones and zeros with positive or negative

charge of electricity - data is lost if there is no source of power

Provides instructions and data to the CPU and stores results of

CPU calculationsinformation constantly changing during


processing

Where Data is Stored


Non-Volatile Magnetic Memory: Tape and Disk
Storage surface coated with magnetic substance
Ones and zeros represented by positive or negative magnetic

polarization

Retains magnetic polarization even without power


Mechanical operation to position a read / write head, over a specific

area of the magnetic surface to:

Write data: write head changes the magnetic pole to positive or negative to
represent a one or zero
Read data: read head senses the positive or negative pole that represents a
one or zero
5

First Magnetic Tape Drive - 1952

Where Data is Stored


Tape
Organizes data sequentially on the tape in the order it receives

the information

More general and simplistic formattingallows tapes written by

one system to be read to a different system in many cases

Cannot directly access each piece of datait reads from the

beginning of the tape until it gets to the data requested

Sequential Access provides good performance to read or write

large amounts of data from start to finish, but very poor


performance for random access

The tape is independent from the tape drive making it easily

portable to other systems or to a safe location

First Magnetic Storage Devices for Computers

First Magnetic Storage Devices for Computers

Where Data is Stored


Disk
Organizes data into specific and addressable areas to read or

write data directly

The disk must be formatted to match the disk addressing

structure of the operating system

Direct access provides fairly consistent performance for mixed

tasks of reading and writing sequential and random groups of


data

Disk performance can be impacted by the length of idle time

necessary to position the read / write head over the area being
addressed

Disk is physically connected to the systemimpractical or

impossible to move the disk to a new location or new system


10

Physical Disk Connections


Rules for physical connection

Type of cable
Number of paths
Physical connectors
Rules for logical connection

HBA

To identify a read or write command vs. data


Format of drive

Addressing scheme

System Bus

Controller

system or
circuit card

ROM
(Read Only Memory)

ESCON
for mainframe
Host bus adapters for open systems
Proprietary cards for AS/400

CPU
MAIN MEMORY
(RAM)

11

How the I/O Works

HBA

Initiating the Read Request

ROM
(Read Only Memory)

System Bus

CPU
MAIN MEMORY
(RAM)

12

How the I/O Works

HBA

Completing the Read Request

ROM
(Read Only Memory)

System Bus

CPU
MAIN MEMORY
(RAM)

13

A Smarter Way to Use Main Memory and CPU

Customer 1
Meter Reading

ROM

Customer 1
Meter Reading
Customer 2
Meter Reading
Customer 3
Meter Reading

System Bus

CPU
CACHE

(Read Only Memory)

HBA

Lets see, customer 1, then customer 2,


what might be next? ... I predict customer 3

Customer 3
Meter Reading

MAIN
MEMORY
(RAM)

Customer 2
Meter Reading
14

Customer 3
Meter Reading

Customer 1
Meter Reading

ROM
(Read Only Memory)

HBA

CPU

CACHE

A Smarter Way to Use Main Memory and CPU

Customer 1
Meter Reading
Customer 2
Meter Reading
Customer 3
Meter Reading

System Bus

CPU
MAIN MEMORY
(RAM)

Customer 2
Meter Reading
15

How the I/O Works

HBA

Initiating the Write Command

ROM
(Read Only Memory)

System Bus

CPU
MAIN MEMORY
(RAM)

16

How the I/O Works

HBA

Completing the Write Command

ROM
(Read Only Memory)

System Bus

CPU
MAIN MEMORY
(RAM)

17

A Smarter Way to Use Main Memory and CPU

ROM
(Read Only Memory)

CACHE

CPU

HBA

The Write Confirmation


is issued as soon as the data
and write command are secure
in a completely fault
tolerant area

WRITE COMMAND
The Customers
Completed monthly Bill

Customer 1
Meter Reading
Customer 2
Meter Reading
Customer 3
Meter Reading

System Bus

CPU
MAIN MEMORY
(RAM)

18

Peripheral Components of a Computer System

ROM
(Read Only Memory)

Tape Drive Device

System Bus

HBA

Network
Router

NIC

HBA

Storage Array

HBA

SAN
Switch

CPU
MAIN MEMORY
(RAM)

19

Data Storage: A Closer Look


Memory Board

Disk Drive
Tape Cartridge

20

The Disk Drive: A Closer Look

21

Formatting the Drive for Direct Access


Track

Cylinder
Sector

The Disk Platter is segmented into a


number of concentric rings, called
Tracks

A specific Track in the same position


on all of the disk platters in a spindle,
together is called a Cylinder

The disk platter is also segmented into


individual wedge shaped sections
called Sectors

A uniquely addressable area within a


disk drive is Cylinder, Head, and Sector
22

Disk Drive Access Time


Seek Time:
The average amount of time
necessary to move the actuator arm
to position the read / write head
over the track

23

Disk Drive Access Time


Latency:
The average amount of time
necessary to wait for the data to
arrive to the read / write as the disk
spins
Also called rotational delay

24

Disk Drive Access Time

Transfer Rate:
The amount of time necessary
to read data from, or write data
to, the platter and move the data
through the disk drive.

25

Disk Drive Performance Variables


Seek time speed
RPM speed of the disk platters
Faster RPM reduces latency
Faster RPM has minor impact on
transfer rate
Disk drive interface speed
Ultra SCSI 40 MB/sec
Fibre channel 100MB/sec

26

Evolution of Disk Technology


Drive capacities continue to increase dramatically from

increased data density

Performance increasing marginally

Increased RPM speed


Increased use of memory and cache at the drive level

Disk drive interfaces driven by industry standards

Ultra SCSI
Fibre Channel
ATA

Industry challenge

Higher capacity per disk drive reduces cost, but


Reduces the number of actuators for a given capacity

27

EMC Storage Offerings


Symmetrix

CLARiiON
CX700 NS700/G

Centera
Centera

DL700

ADIC
Scalar
Series

CX500
CX300

DMX3000-M2
DMX3000

DMX2000-M2
DMX2000

DMX1000

SAN / NAS

Celerra
DMX800 CNS

DMX1000-M2

AX 100

Netwin 110

SAN / NAS /
Backupto-Disk

CAS

Tape &
Tape
Emulation

28

Inside the Disk Arrays


Host Interface

Host Interface

Fault Tolerant
Cache Memory
Array Controller

Array Controller

Disk Directors

Disk Directors

29

RAID Technology

Redundant
Arrays of
Independent
Disks

30

RAID 0: Striping Data Across Many Disks


without Adding Redundancy
Without RAID
3 physical drives
Defined to the host
computer

Raid 0
Defined to the host
computer as above, but
data is physically moved
to balance activity

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 1
Middle

Volume 2
Middle

Volume 3
Middle

Volume 1
End

Volume 2
End

Volume 3
End

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 2
Middle

Volume 3
Middle

Volume 1
Middle

Volume 3
End

Volume 1
End

Volume 2
End

31

RAID 1 or Mirroring

Without RAID
3 physical
drives
Defined to the
host computer

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 1
Middle

Volume 2
Middle

Volume 3
Middle

Volume 1
End

Volume 2
End

Volume 3
End

RAID 1
A mirrored pair
is created for
each physical
volume

Volume 1
Beginning

Volume 1
Beginning

Volume 2
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 3
Beginning

Volume 1
Middle

Volume 1
Middle

Volume 2
Middle

Volume 2
Middle

Volume 3
Middle

Volume 3
Middle

Volume 1
End

Volume 1
End

Volume 2
End

Volume 2
End

Volume 3
End

Volume 3
End

32

RAID 0 + 1 Performance and Redundancy

Without RAID
3 physical
drives
Defined to the
host computer

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 1
Middle

Volume 2
Middle

Volume 3
Middle

Volume 1
End

Volume 2
End

Volume 3
End

RAID 1 + 0
A mirrored pair
is created for
each physical
volume

Volume 1
Beginning

Volume 1
Beginning

Volume 2
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 3
Beginning

Volume 2
Middle

Volume 2
Middle

Volume 3
Middle

Volume 3
Middle

Volume 1
Middle

Volume 1
Middle

Volume 3
End

Volume 3
End

Volume 1
End

Volume 1
End

Volume 2
End

Volume 2
End

33

Data Parity

Parity for 1st


Group = 0

Group 2

Parity for 2nd


Group = 1

Group 3

Parity for 3rd


Group = 1

LOST DATA

Group 1

Group 1

DATA + DATA + DATA = Parity


0 + 1 + 1 = 0

Group 2

1 +

= 1

Group 3

1 +

= 1

34

RAID 5

Without RAID
3 physical
drives
Defined to the
host computer

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Volume 1
Middle

Volume 2
Middle

Volume 3
Middle

Volume 1
End

Volume 2
End

Volume 3
End

RAID 5
A group of
drives are
bound together
as a physical
volume

Volume 1
Beginning

Volume 2
Beginning

Volume 3
Beginning

Parity for
3rd Group

Parity for
2nd Group

Volume 3
Middle

Volume 2
Middle

Volume 1
Middle

Volume 3
End

Volume 1
End

Parity for
1st Group

Volume 2
End

35

RAID Levels
Raid Level

Technique

Application

0
Striping with no Parity
Redundancy

Large Block Performance,

1
Mirrored Disks
Simple Implementation

Highest Availability and Performance

2
Hamming Code
Multiple Check Disks

Large Block Performance


Availability, Poor Cost

3
Striping with Parity
Single Check Disk

Large Block Performance


Availability at Less Cost

Independent Read/Write
Single Parity Disk

Transaction Processing, High Availability,


High Percentage of Reads

Independent Read/Write
Independent Parity Disks

Transaction Processing, High Availability,


High Percentage of Reads

Independent Read/Write
Transaction Processing, High Availability,
Multiple Independent Parity Disks High Percentage of Reads

No

36

Storage Consolidation
Typical customer challenges:

Server/Storage environment

Server/storage islands due to

distributed computing model

Difficult to manage with reduced

manpower

Poor utilization of storage


Integration of infrastructure due to

merger/acquisition is difficult

Asset management is difficult

37

What is a Storage Area Network (SAN)?


A dedicated network carrying block-based storage traffic

Users / Application
Clients

IP
NETWORK

Fibre Channel
NETWORK

LAN
Switches

SAN
Switches
Directors

Servers / Applications

Storage / Application
Data
38

SAN Benefits
High availability and manageability

All servers access same storage


Simplified management
Service for multiple platforms
Application performance

SAN provides a dedicated network


DBMS / transaction processing
Fastest record access
Fast scalability

Hundreds of servers
Hundreds of storage devices
Leverages existing infrastructure
Overcomes distance limitations

Better replication and recovery options


Storage consolidation optimizes TCO
39

What is Network Attach Storage (NAS)?


A network carrying file-based traffic
SAN
Switches
Directors
Gateway
IP
NETWORK

Fibre Channel
NETWORK

LAN
Switches

Users / Application
Clients

Servers / Applications

Storage / File Data


40

NAS Benefits
Global access to information

File sharing
Any distance
Many to one, or one to many
Access from multiple platforms

Consolidation minimizes TCO


Collaboration
Improve time to market
Improve product quality

Information management
Leverage existing security
Leverage existing personnel
Leverage existing infrastructure

Replication and recovery options


Scalable without server changes
41

High Availability
Typical customer issues:
Mission critical data requires 7x24 uptime
No single point of failure (SPOF)
Time to market requirements are tighter
Development cycle is shorter
Development of technology is quicker
Competition is everywhere
Amount of data is growing, backup windows are shrinking
Meet Recovery Point Objectives (RPO) and Recovery Time

Objectives (RTO) easier

42

Data Path Protection

43

What is Path Management Software?


Allows you to manage multiple paths to a device to maximize

application uptime
Path

Refers to the route traveled by I/O between a host and a logical device.
Comprises a host bus adapter (HBA), one or more cables, a switch or hub, an
interface and port, and a logical device.

Multi-Pathing

Configuring multiple paths to a single logical device

Redirect I/O

For Load Balancing


For Path Failover

Monitor

HBAs, Paths, and Devices

Manage

Priorities, Policies for information access


Reconfiguration
Component repair
44

Path Management Overview


Server

4 Paths Configured

2 from each HBA through


switch to each SP

H
B
A
0

Provides Data Access

H
B
A
1

Provides Failover

Upon HBA, switch or SP failure


Provides Potential for Load

Balancing

Four native devices configured

c1t2d1, c1t2d2, c2t2d1 and


c2t2d2

SP A

Mirrored Cache

SP B

Mirrored Cache

45

Local Replication Protection

46

SnapShots: Logical Point-in-Time Views


Pointer-based copy of data

Takes only seconds to create a


complete snapshot
Requires only a fraction of original file
system
Snapshots can be persistent across re-

initialization of the array

Can be used to restore Source data


Up to eight snapshots can be created per

Source LUN

Logical Point-In-Time View

snapshot
snapshot
snapshot
snapshot
snapshot
snapshot
snapshot
snapshot
Snap
Production
Information

47

Snap Shot Copy-on-First-Write


Source LUN
Block A
Production Host

Block B
Updated
Block C
Block D

0
0
0
0
0

0
0
0
0
0

10
0
0
0
0

0
0
0
0
0

Block C in the Reserved LUN


now reflects the change that
the Production Application
made and the pointer is
updated to point to the
Reserved LUN for Block C

Reserved LUN
Original
Block C

SP Memory
Block A
Block B
Block C
Block D

Secondary Host
48

BCVs: Full Image Copies


Physically independent point-in-time copies of

source volume
Available after initial synchronization
Once established, no performance impact
between source / BCV
Can be used to restore or replace source in
event of hardware or software error

Can be incrementally re-established

Only changed BCV data overwritten by source


Up to eight BCVs can be established against a

single source LUN concurrently


Can be any RAID type or drive type
(regardless of source)

Full Image Copies

BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV

Clone
Production
Information
49

Remote Replication Protection

50

Remote Replication Business Drivers


Primary Applications
Disaster Recovery
Business Continuance
Secondary Applications
Backup
Testing
Data Warehousing/mining
Content Distribution
Report Generation

51

Recovery Objectives
Recovery Point Objective (RPO)
How far back in time does data need to be recovered if
disaster occurs ?
Recovery Time Objective (RTO)
How much time will pass after a disaster before operations
are online again?

52

Replication Models
Synchronous
Asynchronous
Periodic incremental update
Traditional asynchronous
Semi-synchronous
Full copy

53

Replication Model - Synchronous


No data exposure
Unit of transfer is individual I/O
Transfer trigger is receipt of I/O from host
No acknowledgement of I/O to host until remote copy is
updated
Write ordering
I/Os are applied to target in the same order they were
received
High bandwidth and low latency are critical
Distance is limited
RPO Zero (No data exposure)
54

Synchronous Model
2
I/O
I/O

ACK
3
4

ACK

1.
2.
3.
4.

I/O from host to local storage system


I/O from local storage system to remote (target) system
Acknowledgement back from remote to local system
Acknowledgement from local storage system back to host

55

Replication Model - Asynchronous Periodic


Update
Unit of transfer is Delta Set
Whats changed since last transfer
Transfer trigger can be discrete event or time cycle
I/O is acknowledged to source host immediately
Write ordering is not an issue
Most recent changes are applied to the destination
Updates are applied atomically
RPO User defined, business driven
Link bandwidth and latency requirements are flexible

56

Asynchronous Periodic Update Model


4

Delta set

Chunk
I/O

ACK
2

ACK

1.
2.
3.
4.
5.

I/O from host to storage system


Acknowledgement from local storage system back to host
Trigger event
Delta Sets from local storage system to remote (target) system
Acknowledgement back from remote to local system

57

Backup/Restore Options

58

Backup to Tape Today


Over 80% of all data backed up today goes to tape
Typical backup operation
Full backups usually done weekly
Incremental - backups of all data since the last backup (full or
incremental)
Differential backups-backup of all data that has changed
since the last full backup
For example, a customer performs a full backup of data on
Sunday but only performs an incremental backup of the data
each the rest of the week
Customers typically keep multiple copies of their

backed up data

Average number of copies is 8.5, some at more than 16:1


EMC Best Practices is 4:1or lower
59

What is Backed up?


Operating Environments

Servers
Desktop PCs
Laptop PCs
Applications

ERP (i.e. SAP, Oracle Apps, Peoplesoft, etc.)


CRM- (i.e. Siebel, etc)
Databases (Oracle, UDB, MS SQL)
Homegrown (using DB technology, file systems, etc.)
Messaging ( Microsoft Exchange, etc)

Application data

For all of the above


Miscellaneous other end user data
Logs and journals

Application transaction logs, database journals, file system journals


60

What is Operationally Restored?


Most restores are at the file and volume level
Frequency ranges from several per day to monthly
Full system restores are rare
Most common restores
Email
Files
Application data

61

EMC Backup-to-Disk Solution: Total Solution with


SAN and / or LAN
Data Center
Use SAN and / or LAN

to centralize all backup

NS600G

Provide higher service

levels

Server with
Backup App

Do full backups and

incrementals to disk

Regularly copy from

SAN
WAN

Tape

disk to tape and move


offsite
Primary Data

Extend the life of

remaining tape
infrastructure

Server with
Backup App

Weekly full
backups
Nightly
Copy from
incremental disk to tape
backups
and move
offsite

CLARiiON
with ATA
62

Operational Backup-to-Disk Process


Application data backed
up to disk
Restores from disk
Data migrated to tape
SAN

Tape

Tape vaulted on / offsite

2
Primary Data

CLARiiON
with ATA
63

Introducing CLARiiON Disk Library


Plug and play

Supports existing backup environment


Appears as tape library (Fibre Channel attached)
Stores data in native tape format
Complete compatibility with existing operation

High performance
Single stream performance up to 80 MB/s
Up to 425 MB/s sustained performance

Application / Backup Hosts

CLARiiON
Disk Library

SAN

Cost effective, reliable, and highly scalable


Data compression (up to 3:1)
Capacity:
DL300: 12.5 TB / 37.5 TB
DL700: 58 TB / 174 TB

Built on proven CLARiiON ATA technology

Creates native tape for offsite storage


Policy-based implementation
Automatic single process for moving data from backup
environment to native tape
Offloads creation of second copy from backup server

Standard
Tape
Library
64

CLARiiON Disk Library Implementation


LAN
Backup
Server

0010011111
1100110010

Backup
Server

0010011111
1100110010

Backup
Server

0010011111
1100110010

SAN
Storage and tape devices
can be connected to the
CLARiiON Disk Library
appliance via Fibre Channel
Emulated tape

000111
110010
01101010111
11010010010
01101010111
11010010010

IMPORT
EXPORT

Physical tape
0010111
1001010

Data flow

CLARiiON
Disk Library

Physical Tape
Libraries / Drives
65

Backup to Disk and Disk Libraries


A New Service-Level Option
HIGH

Disk-to-disk replication
Highest service levels

SERVICE LEVELS

CLARiiON

Highest flexibility
Backup images can be repurposed for testing,
reporting, and more

Tape Library Emulation and backup


CLARiiON
Moderate service levels
Disk Library
Faster backup and recovery than tape

to disk

Moderate flexibility

Tape

LOW

Fastest possible backup and recovery


Least impact to production systems

Disk Libraries support integration with legacy tape


environment
CLARiiON with ATA supports integration with
existing CLARiiON primary storage

Tape Libraries
Lowest service levels
Least flexibility

66

Example: Storage Management


Before

Data gathering from multiple

SAN
?

??

NAS

DAS

sources?
Incomplete, uncorrelated
information?
Different tasks, tools for each
vendor?
Complex and time
consuming?
Making assumptions and
mistakes?

After
Storage Management
Networked
Storage

81%

72%
utilization

78%
utilization

60%

Consistent View of
Storage Environment
Plan and Provision
Monitor and Report
Device Management

utilization

utilization
67

Why Storage Management


Easier to meet service levels

Increase application availability


Expedite problem isolation and resolution
Improve time-to-provision
Helps drive down storage environment costs

Reduce IT staffing costs


Proactive storage and SAN management
Automated provisioning
Current, consistent, correlated information
Increase storage utilization; reclaim capacity

Facilitates compliance with new regulations


Common and consistent storage management information and
processes
68

Enabling ILMEMCs Offering


Information and Content Management
Structured
Information
Management

Enterprise Content Management


Enterprise Document Mgmt
Digital Asset Mgmt
Collaboration

DatabaseXtender

Web Content Mgmt


Records Mgmt/Compliance
ApplicationXtender

Data Movement
Data Migration Tools

Intelligent Data Management

SAN Copy
OnCourse

AVALONidm
DiskXtender Family

Services
Partners

Information
Infrastructure
Management

Protection and Recovery


Remote Replication

Local Replication

SRDF Family
MirrorView
RepliStor
Celerra Replicator

TimeFinder Family
SnapView
Celerra SnapSure

Backup /
Recovery
NetWorker

Availability
PowerPath
AAM
CoStandby
AAdvanced

Storage Management
ControlCenter Family
Visual Family
Replication Manager Family

Tiered Storage
Platforms

ATA

SAN

NAS

CAS

Symmetrix
CLARiiON

Technology

Connectrix

Celerra
NS
Series
NetWin

Centera

Tape
Emulation
CLARiiON
Disk Library

Tape
ADIC
Scalar
Series

69

Summary
After completing this session you should:
Be familiar with the components of a computer and how

they work to retrieve and store data

Be familiar with RAID technology and understand the

different RAID protections

Be familiar with Storage Area Networks and Network

Attached Storage as connection options for storage

Be familiar with the different data protection options that

are available

Understand how all the pieces fit into the Information

Lifecycle Management vision

70

Closing Slide

71

Potrebbero piacerti anche