VMAX3 Introduction
-
Upload
salim-sheik -
Category
Documents
-
view
60 -
download
1
description
Transcript of VMAX3 Introduction
© 2014 IBM CorporationConfidential1
VMAX3
Overview High lights and Operational Considerations December 2014
Brian Ikeda
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential2 [email protected]
Overview VMAX3 FamilyThe VMAX100K, VMAX200K & VMAX400K hereto referred to as VMAX3 Arrays. It can scale from 1 or 2 engines with 2 Systems Bays on the VMAX100K up to 1 or 2 engines with 8 System Bays on the VMAX400K. It comes preconfigured from the factory with the EFD, Performance and Capacity devices configured into necessary resources required to meet customer requirements.
For details see the appendix for data specifications.
Here are some of the high lights
• Dual engines support a maximum of 2 DAE’s of 2.5 (240) or 3.5 (120)
• Systems Bays can be dispersed up to 25 meters from the 1st System Bay.
• 3rd Party Racks are optional.
• Comes preconfigured from factory so onsite configuration is reduced.
• No RAID10, THICK LUN(s). Virtual Provision (VP) for thin devices chunk =1 Track=128 KB.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential3 [email protected]
Overview HyperMax OSHyperMAX OS (Engenuity 5977+) is the software which runs on top of the Dynamic Virtual Matrix to scale out flexibility for cores, cache and hosts interfaces.
An embedded storage hypervisor reduces the hardware and network requirements.
Manages system resources to optimize performance on a wide range of I/O requirements.
Advanced fault tolerances monitoring, detection and correction.
Concurrent maintenance & serviceability features.
Provides function support for host VMAX3 array and for large suite of EMC Storage application software
Define priority of tasks for basic systems maintenance, IO process and application (Data Set Services) processing.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential4 [email protected]
Elements of VMAX3 FAST
Disk Group(s) – Internal physical drives grouped by technology, rotational speed, capacity, form factor and RAID type. Disk groups are
preconfigured and cannot be modified by management software.
Data Pool(s) - Collection of data devices (TDAT) configured from the same physical Disk Group. All Data Pool devices come from the same Disk Group. Data Pool(s) are preconfigured and cannot be
modified by management software.
Storage Pool(s) - Collection of Data Pool(s). Application(s) provision on thin devices in DP(s) in a Storage Pool. This includes all related
TimeFinder/Snap and SRDF/A DSE (Data Set Extension). VMAX3 supports max of 2 Storage Pools per Array. One has to be
marked as a default Storage Pool. Storage Pool(s) are preconfigured and cannot be mofidifed by management software.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential5 [email protected]
Provisioning with FAST
Create port group _.pg as normal
Create initiator group _.ig as normal
Create host addressable TDEV(s) using Solutions Enabler 8.0.1 or Unisphere 8.0.1
Create FAST Managed Storage group (SG) from Storage Resource Pool (SRP) symsg –name __.sg –srp SP1
Add Host addressable TDEV(s) to FAST Managed SG
symsg –name __.sg add tdev1:tdev2
Create view as normal using new SG.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential6 [email protected]
Storage Group (SG) FAST Provisioning
Changes provisioning storage from current FAST VP
New command from
symsg -sid ssid
create SRGNAME_sg
-slo (Service Level Objective)
-srp (Storage Resource Pool)
-wl (Work load)
Devices can be in 1 FAST Managed SG only to prevent mangement conflicts
All devices within SG have to have same SRP, SLO.
If only –SRP specified without –SLO or –WL, then SLO= Optimized (default
Parent SG cannot be FAST managed. See documentation on cascading.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential7 [email protected]
Defined SLO (Service Level Objectives)
SLO Flash 15K 10K 7K ELT RT Use
Diamond Alloc .8 ms HPC
Platinum Alloc 3.0 ms High OLTP
Gold Alloc 5.0 ms Heavy IO, DB Logs
Silver Alloc 8.0 ms DB, Virtual Apps
Bronze Alloc 14.0 ms Backup, archive
Optimized(default)
Alloc N/A Places most active on High and least active on lower cost
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential8 [email protected]
Defined Workload
WorkLoad Description
OLTP Small block IO workload
OLTP with Replication
Small block IO workload with local or remote replication
DSS Large block IO workload
DSS with Replication Large block IO workload with local or remote replication
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential9 [email protected]
Embedded NAS (eNAS)
eNAS enables virtual instances (VM) of VNX Software Data Movers and Control Stations. See Appendix VMAX3 Family Data Specification for more information.
VMAX3 cannot be upgrade to eNAS after initial install.
File support only, no block support.
750 TB of useable file systems on VMAX3
Maximum of 16 TB for individual file system on VMAX3
EMC provides a one time setup as follows:
• 2 Control Stations VM
• 2 (default and max) Software Data Movers – VMAX100K (1 Active & Standby)
• 2 (default and 4 max) Software Data Movers – VMAX200,400K (3 Act & 1 Standby)
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential10 [email protected]
TIMEFINDER SNAPVX
HyperMAX (5977) supports TIMEFINDER SnapVX
Local replication to create point in time (snapshots) of critical data.
SnapVX creates snapshot by storing changed tracks in SRP of source device.
TimeFinder SnapVX doesn’t require a targt device or a source/target pair
New fields created SnapVX Source & Target
To access point in time copy, create link with snapshot data to a host mapped target device
Provides for efficient usage of metadata and flash memory because target devices is not required when snapshot is created.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential11 [email protected]
ProtectPoint
Protect Point runs as a service (application) on the VMAX3 and can communicate directly to Data Domain.
GA 1st Quarter 2015
For more high level details see Appendix ProtectPoint Backup & Restore
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential12 [email protected]
Summary
VMAX3 documentation found on the EMC Support site is being added on an ongoing basis.
Of the available VMAX3 documentation, there are references to features, tools which are not available or documented on EMC Support site.
VMAX3 Preconfiguration will need to be evaluated on a case by case basis, since this is performed through the client/EMC order process.
Hardware upgrades to onsite VMAX3 may be problematic. See Appendix Release Notes for serviceability.
Preconfigured Data Pools within Storage Resource Pools are thin and will need to be examined for oversubscription.
SnapVX will have to be tested, examined for suitability as needed. Regular TF Snap,Close, & Mirror are still supported.
SRDF between VMAX3 (5977) and VMAX (5976-) will have to be evaluated for support
ProtectPoint will have to examined when it becomes available 1st Qtr 2015.
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential13 [email protected]
Appendix VMAX3 Family Data Specification
EMC VMAX3 Family Data Specification Sheets
VMAX100K, VMAX200K & VMAX400K
VMAX3 Family Data Spec Sheet
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential14 [email protected]
Appendix ProtectPoint Backup
AAAAA
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential15 [email protected]
Appendix ProtectPoint Restore
AAAAA
VMAX3 Highlights & Operational Considerations
© 2014 IBM CorporationConfidential16 [email protected]
Appendix Release Notes
AAAAA
VMAX3 Releast Notes 12-4-2014