You are on page 1of 5

Clariion concepts Models: CX 300 & CX3-20/40/80 Tasks: Provisioning & deprovisioning the storage.

age. Administration through Navisphere Manager. Creating luns, raid groups, and storage groups. Coordinating with the EMC in case of any hardware failures.

Clariion Hardware: 1. 2. 3. 4. 5. 6. 7. Building blocks : DAE,DAE2P,DPE,SPE,SPS Enclosures will have 10/15 disks based on the model. 2 storage processors, CMI (Fiber channel) Link control cards. Front end ports & Back end ports. Each storage processor will have 2 cpus Dual power supply & stand by power supplies.

Clariion storage concepts 1. 2. 3. 4. Max 4GB cache/SP & Mirrored write cache between SPs Flare software is installed on the 1st 4 disks. Flare does provisioning, resource allocation, process scheduling... etc... Cache page size 8KB. & cache is not user configurable. Read & write caches are user configurable. 5. Cache Flushing techniques: Idle cache, watermark cache & Forced cache flushing. 6. If data is written to SP-A cache, same data is transferred to the SP-B cache through CMI. 7. Page sizes: 2, 4, 8,16KB. Exchange 4KB, SQL 8KB, Oracle 16KB page sizes. 8. Vault is 1st 5 disks (CX) or 9 disks (FC series), write cache data is stored in vault when power failure. Raid3 is implemented on. 9. PSM- persistent storage manager, records configuration information. Created during initialization of array, resides in 1st loop of drives... Size 1GB CX & 512 MB on FC. Available in the vault drives 10. Default element size is 64K 11. 128 luns/Raid group & 2048 luns/array. 12. Snap view Local replication ( Snapshot- point in time views & clonesPoint in time copies) 13. Snapshot roll back feature provides instant restore to source volume(20% of the Source volume) 14. Reserved Lun pool is created for each sp, and it holds the all original data of the source Lun when the host writes to a chunk for the 1st time. 15. Clone is full copy of the source volume. 16. Private Lun must be created before any clone operation, itll have fracture log, modified tracks b/w source & Lun.

17. Mirror view/s ensures that there is an exact byte for byte copy at both local & remote Clariions. 18. Mirror view/A duplicate changes in production site data to secondary site at regular intervals. 19. Mirror view states: Out-of sync/in-sync/Consistent/synchronizing/rolling back (MV/A only). 20. Fracture log captures pending writes when link is down & resyncs when link is up since the time of link failure. 21. Write Intent log: will have pending write during the link failure, and sent to the remote site when link is up to regain the synchronous state. 22. SAN copy is used for data replication between 3rd party storage. 23. Access Logix provides Lun masking feature that disables unwanted access from the other servers. 24. Supporting Raid types: Raid1, 1/0 3, 5. Global hot spares 25. Host agent (naviagent.exe) registers the servers HBA with the attached storage system when the agent service starts, this action sends the initiator records for each HBA to the storage system and sends drive mapping information to the attached clariion storage. 26. There are four failover modes in Clariion. a) Failover mode 0: Lun Based trespass mode, works with Auto trespass feature; whenever Lun is trespassed a Unit attention message is recorded. A host with no failover software should use the combination of failover mode 0 & Auto-trespass disabled. b) Failover mode 1: Passive not ready mode, this mode is most commonly used with the power path. c) Failover mode 2: DMP mode, works with VERITAS DMP,this mode is similar to the failover mode 0 auto-trespass enabled, but Unit attention messages are suppressed. d) Failover mode 3: Passive always ready mode, An I/O request sent to the non owning SP will be rejected, this mode is only used on AIX server under very specific configuration parameters.

27. Initiator records are used to control access to storage -system data. 28. Admhost utility works with the SANcopy, snap view, Mirror view/A devices. 29. Use admhost command to scan for new storage devices, assign drive letters to the devices, mask storage devices inaccessible to the server OS, flush cache data to disk, list drive letters& worldwide names for all server devices. 30. Admhost utility runs on servers attached to CX series& Ax series storage systems. 31. Reserved IPs for Storage processors by EMC 192.168.1.1/2 & 128.221.1.250/251.

32. Registering HBA in Clariion Windows: go to services window, stop & start service Navisphere Agent Aix : rc.agent stop/start HP-UX : #/sbin/init.d/agent stop/start Linux: #/etc/init.d/naviagent stop/start Solaris : #/etc/init.d/naviagent stop/start VMware : #/etc/init.d/naviagent stop/start Netware : on netware host restart the Navisphere agent, In netware server console screen enter sys:emcagentnavagent.nlm -f sys:emcagentagent.cfg

33. #agentID.txt --ensures that the Navisphere agent binds to the correct HBA for registration and registers the host with the correct storage system. The text file must contain a) full qualified hostname of the host b) IP address of HBA port that you want naviagent to use. Line1: host28.mydomain.com & Line2: 192.111.222.3 A file resides in this path:

Windows: C:/programfiles/EMC/Navisphere Agent VMware :/var/log/ HostIdFile.txt AIX: /etc/log/HostIdFile.txt Hp-UX : /etc/log/HostIdFile.txt Linux: /var/log/HostIdFile.txt Solaris: /etc/log/HostIdFile.txt

SP Collects: The following data that is collected using the SP Collects from both the SPs:

Ktdump Log files iSCSI data FBI data (used to troubleshoot backend issues) Array data (sp log files, migration info, flare code, sniffer, memory, host side data, flare debug data, meta Lun data, prom data, drive meta data, etc) PSM data

RTP data (mirrors, snaps, clones, SANcopy, etc) Event data (windows security, application and system event files) LCC data Nav data (Navisphere related data)

FLARE: Fibre Logic Array Runtime Environment. Flare is loaded in 1st 5 drives ( vault disks) Do not configure any user-host Lun space on 1st five drives. 6 GB pre disk is used. 1st 5 drives will have PSM Lun, Flare database Lun, vault data. Vault--save area for write cache when catastrophic failure of SP. Engineering mode ctrl+shift+F12 Drives 0 and 2 are mirrored for SPA. Drives 1 and 3 are mirrored for SPB. Drive 4 contains space for Core Dump. Stripe across drives 0, 1, 2 contain the PSM. Another Stripe across drives 0, 1, 2 contains the Database. Stripe across drives 0, 1, 2, 3 contain reserved space. Stripe across drives 0, 1, 2, 3, 4 contain the Write Cache Vault.

CX3 comparison

CX-4 updates:

Models: CX4-120/240/480/960 & Cache: 6GB/8GB/16GB/32GB Ultra flex I/O modules (FC & iSCSI) SFP: small form factor pluggable-optical Fc ports speeds are in 1/2/4Gbps ISCSI ports speeds are in 10/100/1000Mb/s 4FE version: That provides four front-end optical ports; 2FE/2BE version: That provides two front-end optical ports and 2 copper backend ports; 4BE version: that provides four back-end ports 2FE/1BE version:

HBA replacement steps Login in to the MSEDSBSCDS001 Obtain the new WWN of the HBA card Save the current config Select the alias for the server "MSESAUXCDSN002_HBA0A" Replace the old wwn with the new wwn Save the changes and enable config Please make a note the MSESAUXCDS002 is connected to Port5 on both the switches. After zoning In clariion deregister the old wwn's of the server, in the connectivity window. Then start & stop the naviagent on server. # /etc/init.d/agent stop # /etc/init.d/agent start New wwn's should display with server name. After that go to storage group and add host again. Reboot the server if all disks are not able to see.

You might also like