Uploaded by Sinisa Antolic

HPE 3PAR StoreServ Simulator 3.2.2

advertisement
HPE 3PAR StoreServ
Simulator
Changes to object, file, storage management and
reporting
Contents
HPE 3PAR StoreServ Simulator.......................................................................................... 2
Overview........................................................................................................................ 2
Supported Platform......................................................................................................... 2
Supported Features........................................................................................................ 2
Simulator Installation......................................................................................................... 2
Minimum System Requirements.....................................................................................2
Installation and Configuration Summary.........................................................................3
VMware ESXi Server Installation........................................................................................3
Installing the Cluster Nodes............................................................................................ 3
Installation of the Enclosure Node..................................................................................4
Network Configuration.................................................................................................... 4
VMware Workstation.......................................................................................................... 6
Installation of the Cluster Nodes Overview.....................................................................6
Enclosure Node Installation............................................................................................ 6
Network Configuration using VMware Workstation.........................................................7
Technical white paper
Technical white paper
Page 2
HPE 3PAR StoreServ Simulator
Overview
The 3PAR Simulator software provides a virtualized HPE 3PAR StoreServ Storage System without the
corresponding physical storage hardware requirements. The features normally provided by the hardware
components, including the ASIC, HBA and enclosure are emulated in software. The 3PAR Simulator
provides the same look, feel and primary interaction capabilities of a physical storage system including
support for management, administration and reporting through the HP 3PAR StoreServ Management
Console (SSMC), remote CLI, and client software access to the CIM-API and WS-API interfaces.
Supported Platform
The 3PAR Simulator software release version 3.2.2 MU4 supports the configuration of up to 2 simultaneous
instances, differentiated by name and serial number, of a 2-node HPE 3PAR StoreServ 8200 Storage
System running HP 3PAR OS version 3.2.2 MU4.
Supported Features
The following 3PAR StoreServ Storage System features are supported:
•
Up to 48 HDDs, 4 cage configuration
•
Cage types – DCN2, DCS7, DCS8
•
Disk types – FC, SSD, NL
•
3PAR Management Console support
•
CLI and Remote CLI support
•
CIM-API (SMI-S) and WS-API support
•
Storage Provisioning including Thin-Provisioning
•
Exporting Virtual Volumes
•
Adaptive Optimization (AO)
•
Dynamic Optimization (DO)
•
Local Replication (Snapshot)
Remote Replication (RCIP) - requires 2 instances of the simulator (Requires more resources)
Important to note – When using the HPE 3PAR Simulator with Remote Copy, extra resource requirements will be necessary for the VM’s which
make up the simulator instance.
NOTE: the license validity for HPE Storeserv 3PAR Simulator has been extended till
December 2024.
Simulator Installation
Installation of the HPE 3PAR simulator requires a VMware environment. Requirements include

One VMware VM assigned as the ESD node

Two VMware VM’s assigned as cluster node 1 and cluster node 2
Minimum System Requirements
The 3PAR Simulator requires a minimum of VMware ESXi 5.5 or above, or VMware Workstation 11 or later. The simulator configuration requires
deploying three VMs; two VMs simulating the cluster nodes, a third VM to simulate the enclosure and private network configuration to enable
communication between the three VMs. The simulator has not been tested on any other variations of virtual machines and it is not supported in
Technical white paper
Page 3
any other configuration that tested. It should be noted that the user can add additional resources but is not required unless simulating a RCIP
configuration as noted below.
The minimum system resources required for each VM are:

One virtual CPU (single core)

Three virtual NICs

2GB RAM (4GB is required for use in a RCIP configuration)

50GB disk space (full or thin-provisioned)
Installation and Configuration Summary
Setup of the 3PAR Simulator configuration will require completion of the following steps:
1.
2.
3.
4.
5.
Cluster Nodes Installation – Involves installation of two VMs (sim_node0 & sim_node1) using the cluster node OVF package.
Enclosure Node Installation – Involves installing one VM (sim_esd) using the enclosure node OVF package.
Network Configuration – Involves creating a private network for communication between the three VMs. The network configuration varies for the
VMware ESXi Server and VMware Workstation setup, follow the instructions in the corresponding section of this guide.
Simulator configuration and bring up – Involves configuration of the cluster node VMs, enclosure node VM and bringing them all up to form the
storage system simulator.
Installation of the HPE 3PAR StoreServ Management Console (SSMC) 2.3 and 3PAR CLI application v3.2.2 MU4 (optional).
VMware ESXi Server Installation
Installing the Cluster Nodes
The cluster node (clus_node) package is included in the HP-3PAR-StoreServ-Simulator-3.2.2MU4.zip file. Extract the files to a folder on your
computer. The uncompressed file size of the clus_node package is ~ 6.4GB
Note
Each "Deploy with OVF template" operation in vSphere Client can take up to 30 minutes to complete with high speed LAN networking. It is highly
recommended that the installation is performed from a high speed network or local disk.
The cluster node installation steps are as follows:
Step 1: Unzip the cluster node OVF package
There are 3 VM files in the cluster node OVF package:
•
clus_node_template_3.2.2.MU4_field_mode.ovf
•
clus_node_template_3.2.2.MU4_field_mode -disk1.vmdk
•
clus_node_template_3.2.2.MU4_field_mode.mf
Step 2: Create a VM (sim_node0) using the "Deploy OVF template" option in vSphere
•
Login to the VMware vSphere Client
•
Click the "File" option at the top menu
•
Click "Deploy OVF template"
•
In the "Deploy from a file or URL" dialog, specify the full pathname of the clus_node_template_3.2.2.MU4.ovf file
•
Click "Next" - shows the OVF Template details
•
Click "Next" - type in the name of the new VM
Example: sim_node0
•
Click "Next" - shows the disk format, choose thick or thin provisioning
•
In “Network Mapping” screen, select VM Network
Technical white paper
•
•
Page 4
Click "Next" - Ready to complete?
Click "Finish" - Deploying Your new VM
Step 3: Create a second VM (sim_node1) using the vCenter “clone” or “Deploy OVF template” option in vSphere
You can use the "clone" feature in vCenter to create an additional VM that is identical to the one that was created in Step 2 above, and provide a name
for the cloned VM (Example: sim_node1). If you are not using vCenter, you can create the second VM by following the steps below:
•
Login to the VMware vSphere Client
•
Click the "File" option at the top menu
•
Click "Deploy OVF template"
•
In the "Deploy from a file or URL" dialog, specify the full pathname of the clus_node_template_3.2.2.MU4_field_mode.ovf file
•
Click "Next" - shows the OVF Template details
•
Click "Next" - type in the name of the new VM
Example: sim_node1
•
Click "Next" - shows the disk format, choose thick or thin provisioning
•
In “Network Mapping” screen, select VM Network
•
Click "Next" - Ready to complete?
•
Click "Finish" - Deploying Your new VM
Installation of the Enclosure Node
The enclosure node (esd_node_template_3.2.2.MU4) OVF package is included in the HP-3PAR-StoreServ-Simulator-3.2.2 MU4.zip file. Extract the
files to a folder on your computer. The uncompressed file size of the enclosure node package is ~ 460MB.
Note
The enclosure node installation steps for creating the third VM (esd_node) are similar to the cluster node installation except that the
esd_node_template_3.2.2 MU4 OVF package must be used
Step 1: Unzip the enclosure node OVF package
There are 3 VM files in the enclosure node OVF package:
•
esd_node_template_3.2.2.MU4.ovf
•
esd_node_template_3.2.2.MU4 -disk1.vmdk
•
esd_node_template_3.2.2.MU4.mf
Step 2: Create VM (sim_esd) using the "Deploy OVF template" option
•
Login to the VMware vSphere Client
•
Click the "File" option at the top menu
•
Click "Deploy OVF template"
•
In "Deploy from a file or URL", specify the full pathname of the esd_node_template_3.2.2.MU4.ovf file
•
Click "Next" - shows the OVF Template details
•
Click "Next" - type in the name of the new VM
Example: sim_esd
•
Click "Next" - shows the disk format, choose thick or thin provisioning
•
In “Network Mapping” screen, pick VM Network
•
Click "Next" - Ready to complete?
•
Click "Finish" - Deploying Your new VM
Technical white paper
Page 5
Network Configuration
Once the VMs have been deployed, the network configuration must be completed to connect them to the appropriate networks. The VMs are
configured with three vNICs and the network label is set to VM Network by default for all the three vNICs. The default configuration must be changed
as follows:
•
The first vNIC will be dedicated for use with management tools. It must be connected to a network allowing communication to the servers
running the HP 3PAR Management Console, remote CLI or client software accessing the CIM-API (SMI-S) or WS-API. Once configured, these tools
will connect to the 3PAR Simulator via the cluster IP address specified during the “Out of the Box” (OOTB) procedure.
•
(Optional) The second vNIC will be dedicated for use by Remote Copy if you plan on configuring two separate instances of the 3PAR Simulator to
leverage this capability. If Remote Copy is configured this interface must be connected to the Remote Copy network. This applies only to the
cluster nodes (sim_node0 & sim_node1) as the enclosure node (sim_esd) is not part of the Remote Copy network. The second vNIC in the
enclosure node (sim_esd) is unused and the network label can be left to the default VM Network.
•
The third vNIC must be connected to a private network connecting all three VMs (sim_node0, sim_node1 & sim_esd) with a virtual switch
(vSwitch). The virtual switch should not be connected to a physical adapter as the traffic will only be routed locally (assuming the VMs hosted on
the same ESXi) to allow communication between the three VMs in the 3PAR Simulator configuration. Follow the instructions below in Step 1:
vSwitch configuration and Step 2: Connect the third vNIC in each VM to the Private Network to set up the private network.
Step 1: vSwitch configuration
To configure a vSwitch:
•
Log in to the vSphere client and select your ESX host from the inventory panel
•
Click the “Configuration” tab in the right side top panel
•
Click on “Networking” in the Hardware frame
•
Click on “Add Networking” (top right corner)
•
Accept the default connection type, “Virtual Machine” and click “Next”
•
Accept the default “Create a vSphere standard switch”
•
Deselect the network adapter(s).
•
Example: If the adapter vmnic2 is selected, deselect it. (See Figure 1 below)
Figure 1 Sample vSwitch configuration with all adapters unselected
•
•
•
Click “Next”
Accept the default Network Label and VLAN ID None (0). Make note of the Network Label, this label must be selected in the next step to connect
the vNIC to the correct network
Click “Next” and then “Finish” to complete the vSwitch configuration (See Figure 2 below)
Figure 2 Sample vSwitch configuration
Step 2: Connect the third vNIC in each VM to the Private Network
Connect the third vNIC in each VM to the private network corresponding to the vSwitch created in the previous step.
To connect the vNIC:
•
Log in to the vSphere client and select the sim_esd VM from the inventory panel
•
Click the “Getting Started” tab (in the right page)
•
Click on “Edit virtual machine settings”
•
Click on the network interface “Network Adapter 3”.
Technical white paper
•
•
•
Page 6
In the Network label pull down menu, select the network label created in the previous step.
Example: vSwitch1 was created in the previous step and the network is VM Network 2
Select “VM Network 2”
Click “OK”
Figure 3 Sample Private Network configuration
VMware Workstation
This section describes the HPE 3PAR StoreServ simulator installation and network configuration procedures
for VMware Workstation 11 and above.
Installation of the Cluster Nodes Overview
The cluster node (clus_node) package is included in the HP-3PAR-StoreServ-Simulator-3.2.2 MU4.zip file.
Extract the files to a folder on your computer.
Step 1: Unzip the clus_node OVF package
There are 3 VM files in the clus_node OVF package:
1. clus_node_template_3.2.2.MU4_field_mode.ovf
2. clus_node_template_3.2.2.MU4_field_mode -disk1.vmdk
3. clus_node_template_3.2.2.MU4_field_mode.mf
Step 2: Create a VM (sim_node0) using the "Open" option within VMware Workstation
1. Launch the VMware Workstation
2. Click the "File" option at the top menu
3. Click "Open"
4. In "Open file or URL", specify the full pathname of the clus_node_template_3.2.2.MU4_field_mode.ovf file
5. Click "Open" - shows Import Virtual Machine
6. Enter a name of the node
Example: sim_node0
You can leave the default path for the location of the VM
7. Click “Import”
Step 3: Create a second VM (sim_node1) using the "Open" option within VMware Workstation
1. Launch the VMware Workstation
2. Click the "File" option at the top menu
3. Click "Open"
4. In "Open file or URL", specify the full pathname of the clus_node_template_3.2.2.MU4_field_mode.ovf file
5. Click "Open" - shows Import Virtual Machine
6. Enter a name of the node
7. Example: sim_node1
8. You can leave the default path for the location of the VM
9. Click “Import”
Technical white paper
Page 7
Enclosure Node Installation
Step 1: Create a VM (sim_esd) using the "Open" option within VMware Workstation
1. Click the "File" option at the top menu
2. Click "Open"
3. In "Open file or URL", specify the full pathname of the esd_node_template_3.2.2.MU4.ovf file
4. Click "Open" - shows Import Virtual Machine
5. Enter a name of the node
Example: esd_node
You can leave the default path for the location of the VM
6. Click “Import”
Network Configuration using VMware Workstation
Once the VMs are deployed, the network configuration must be done to connect them to appropriate networks. The VMs are configured with three
vNICs and the network label is set to VM Network by default for all the three vNICs. The configuration must be changed as follows:
•
The first vNIC (Network Adapter in g
fi ure 4) will be dedicated for use with management tools. It must be connected to a network allowing
communication to the servers running the HP 3PAR StoreServ Management Console, remote CLI or client software accessing the CIM-API (SMI-S)
or WS-API. Once configured, these tools will connect to the 3PAR Simulator via the cluster IP address specified during the “Out of the Box” (OOTB)
procedure.
•
(Optional) The second vNIC (Network Adapter 2 in figure 4) will be dedicated for use by Remote Copy if you plan on configuring two
separate instances of the 3PAR Simulator to leverage this capability. If Remote Copy is configured this interface must be connected to the
Remote Copy network. This applies only to the cluster nodes (sim_node0 & sim_node1) as the enclosure node (sim_esd) is not part of the
Remote Copy network. The second vNIC in the enclosure node (sim_esd) is unused and the network label can be left to the default VM Network.
•
The third vNIC (Network Adapter3 in figure 4) must be connected to a private network connecting all three VMs (sim_node0, sim_node1 &
sim_esd) with a Lan Segment for traffic to be routed locally, allowing communication between the three VMs in the 3PAR Simulator
configuration. Follow the instructions below in Step 1: Configuring a LAN Segment and connecting vNIC to the LAN Segment Network to set up the
private network.
Step 1: Configuring a LAN Segment
To configure a LAN Segment:
1. Launch VMware Workstation
2. Right-Click on sim_esd, select “Settings”
3. Click on “Networking Adapter 3” in the Hardware frame
4. In the Network Connection frame, select LAN Segment
5. Click on the tab below that says “LAN segment”. A global LAN Segments Window will pop up
6. Click “Add” and enter a name for the LAN Segment
Example: Simulator
7. Ensure that the Network Adapter 3 is connected to the LAN Segment you just created by selecting it using the drop down menu (See Figure 4
below)
8. Click “OK”
Technical white paper
Page 8
Figure 4 Virtual Machine Settings - LAN segment selection for Network Adaptor 3
Important
Repeat the above steps to connect Network Adapter 3 for both sim_node0 & sim_node1 to the LAN segment
Step 2: Configuring a management network.
In order to use the HP 3PAR StoreServ Management Console or 3PAR remote CLI on the system that is running VMware Workstation and hosting the
3PAR Simulator instance, or another VM instance you will need to configure Network Adapter 1.
1.
2.
3.
4.
5.
6.
7.
8.
9.
Right-Click on sim_node0, select “Settings”
Click on “Network Adapter” in the Hardware frame
In the Network connection frame, select “Host-only”
Click “OK”
Click the "Edit" option at the top menu
Click "Virtual Network Editor”
Select the host only type VMnet
Note the Subnet Address and the Subnet Mask. You will need this later during the OOTB process when configuring the node IP address
that talks to the StoreServ Management Console, CLI Utility, Putty, etc.
After creating the “Host Only Network” you may need to reboot your machine in order to initialize the network stack. Few users running VMware
Workstation 10 have reported that is the only way they could get the network to work in their case.
Technical white paper
Page 9
Figure 5 Virtual Machine Settings - Host only selection for Network Adaptor 1
Important
During the “Out Of The Box” (OOTB) procedure, you will need to provide a Static IP and the same subnet mask to be able to communicate with the HP
3PAR Simulator storage system instance form the 3PAR StoreServ Management Console on your laptop or on a separate VM as well. Please refer to
the SSMC documentation for resources requirement
Note
Note about VMware Networking: When you assign a Host-Only network profile to a specific VMNetx , VMware Workstation will
automatically create a new Class C Subnet for that VMnet (i.e. 192.168.x.x) Every installation is different, so it is important to check
what subnet is actually in use on your machine using the VMware Networking Editor Utility (under Edit on the top menu of VMware
Workstation home page). This network will by default have DHCP turned on (assigning addresses from .128 through .254), as well as
a default gateway of .1 on that subnet.
In most cases, VMware Workstation install would create VMNet1 interface on your system. The default IP assigned to the VMnet1 interface is:
192.168.80.1. THIS IS NOT AN IP ADDRESS YOU ASSIGN TO THE 3PAR SIMULATOR.
Example:
In this case the IP address of 192,168.80.10 and subnet mask of 255.255.255.0 have been assigned to the HP 3PAR Simulator node during OOTB
procedure.
Technical white paper
10
Figure 6 Host IP Address Settings - ipconfig for VMNet1
Figure 7 VMWare Virtual Network Editor - Host only IP address configuration for VMnet 1
Simulator Setup
After node installation and network configuration are complete, power up all VM’s.
Page
Technical white paper
Page
11
Step 1: Start the esd daemon
Log in to the VM deployed with the esd package with user root password root
Run createcages to create the desired cage configuration.
The createcages utility will display the following menu:
Choose from the menu:
1. 1 cage, 24 HDDs
1 DCN2 cage with 24 FC drives
2. 4 cages, 96 HDDs
1 DCN2 cage with 12 FC drives and 12 SSD drives
2 DCS8 cages with 12 FC drives and 12 SSD drives each
1 DCS7 cage with 12 NL drives and 12 SSD drives
3. 4 cages, 48 HDDs
1 DCN2 cage with 12 FC drives
3 DCS8 cages with 12 FC drives each
4. 4 cages, 96 HDDs
1 DCN2 cage with 12 FC drives and 12 SSD drives
3 DCS8 cages with 12 FC drives and 12 SSD drives each
5. exit (you can create cages using esd)
Select a number(1-5) to create cages and then exit
Once you have the HDW configured, you must start the esd daemon by running the command esd.
See the section Error: Reference source not found for more information on how to manage cages & disks using esd.
Step 2: Simulator Node Configuration
1.
2.
3.
Log in to the first node as user console password crH72mkr
Choose: Option 1 (Simulator Cluster Node Configuration)
a) Enter 0 for the node ID
b) Choose one the serial # to assign to the cluster node ( 1699678 or 1699679 )
c) Confirm
Log in to the second node as user console password crH72mkr
a) Enter 1 for the node ID
b) Assign the same serial number to this node as the one you assigned 2-b ( 1699678 or 1699679 )
c) Confirm ‘y’
Important
•
The serial number must be the same for both nodes in a simulator instance
•
One of the two serial numbers must be used in order to activate the preloaded licenses. Failing to use one the serial numbers will leaded to
an instance where no license is applied and you would have to re-install.
4.
5.
6.
7.
8.
Choose: Option 5 in each VM to reboot
Once the VMs are booted, log in to the first node as user console password crH72mkr
Choose: Option 2 to run the Out Of The Box procedure
Enter “yes” to continue
Follow the prompts to finish the configuration using the information you collected in the Simulator Setup section.
When OOTB completes, the HP 3PAR Simulator storage system cluster node configuration is done.
See the section Error: Reference source not found for more information on how to manage cages & disks using esd.
Step 3: Verifying the configuration
1. Open the HP 3PAR Management Console
2. Connect to the cluster node IP address
3. Login with user 3paradm password PTzk5fVt
Technical white paper
Page
12
4.
Check the controller nodes, cages and disk configuration.
Simulator Best Practice
1.
2.
3.
Unless the entire storage system (cluster nodes and esd node) is being brought down or restarted, do not terminate the esd daemon alone
while the nodes are up and running. Terminating the esd daemon has the effect of disks going offline leading to TOC quorum loss and
eventually the nodes go into power fail mode.
Gracefully shutdown the HP 3PAR Simulator by first shutting down the cluster nodes, then stop the esd daemon and shutdown the esd VM.
Use the following steps:
A. Login to node 0 as user 3paradm and password PTzk5fVt
B. Issue the command
shutdownsys halt
C. Once the cluster node VMs (node 0 and node 1) are powered down:
i. login to esd VM as user root password root
ii. Issue the command
esd stop
D. Once the daemon is stopped, issue the command
shutdown –h 0
When powering back up the HP 3PAR Simulator you need to:
A. Bring the esd VM first and start the esd daemon
B. Then power up both the cluster nodes.
Enclosure and Cage Configuration
Start the Daemon
You can start the daemon by running either of the following commands:
esd
esd start
esd data directory: /common/esd/.esddata
The esd maps each disk to a file. By default, all data files are created in a directory called .esddata under /common/esd.
You can override the default parent directory /common/esd when you start esd using the –l option:
esd start –l <new-dir>
Stop the Daemon
You can stop the daemon by running the following command:
esd stop
Note
You cannot start esd again without stopping it first as only one instance of esd can be running at any time
Cage WWN and Disk WWN
Some of the subcommands require cage WWN or disk WWN to perform the requested operation. The esd automatically generates cage WWN and
disk WWN when a new cage or disk is added. You can use the list command described later in the section to get the WWNs.
Add a Cage
Synopsis:
esd add cage [mag-num=<num> | disk-num=<num>] cage-type={DCN2|DCS7|DCS8}] [disk-type={FC|SSD|NL}] node=<node0-name>
[node=<node1-name>]
mag-num
specifies the number of magazines to create, default 24 magazines
disk-num
specifies the number of disks to create, default 24 disks
cage-type
specifies the type of cage to create: DCN2, DCS7, DCS8
disk-typespecifies the type of disks to create: FC, SSD, NL
node
specifies the nodes to connect
Technical white paper
Page
13
Important
1.
2.
3.
4.
5.
For the first cage configuration, the parameter cage-type is optional. By default the first cage created will be DCN2 with FC drives. A user is
allowed to override the default by specifying the cage-type and/or disk-type. The parameter cage-type is mandatory for additional cages.
The version 1.2 release of HP 3PAR Simulator supports only a 2-node cluster, so you can configure only one DCN2 cage with the simulator. If
you need to configure more than one cage, use cage types DCS7 or DCS8.
If the parameter disk-type is not specified, the following defaults will apply for each cage type:
a. DCN2 and DCS8 cages: FC
b. DCS7 cage: NL
If the parameter disk-num or mag-num is not specified, the default number is 24.
The simulator supports up to 4 cages with at most 48 disks. For example, a configuration with 2 cages and 24 disks per cage or 4 cages with
12 disks per cage has reached the limit. Additional cages/disks cannot be added until some disks are deleted.
Examples:
1.
Add the first cage with the default configuration:
# esd add cage node=vm-a node=vm-b
Note
The above command adds a DCN2 cage with 24 FC drives, connecting to cluster nodes vm-a and vm-b
2.
Add the first cage or additional cage by specifying cage-type:
# esd add cage cage-type=DCS8 node=vm-a node=vm-b
Note
The above command adds a DCS8 cage with 24 FC drives, connecting to cluster nodes vm-a and vm-b
3.
Add an additional cage by specifying cage-type, disk-type, and disk-num:
# esd add cage cage-type=DCS8 disk-type=SSD disk-num=12 node=vm-a node=vm-b
Note
The above command adds a DCS8 cage with 12 SSD drives, connecting cluster nodes vm-a and vm-b
Add Magazine(s)
Synopsis:
esd add mag -w <cage-wwn> [disk-type={FC|SSD|NL}] [mag-num=<num>]
-w <cage-wwn> specifies the associated cage to add the magazine
mag-num
specifies the number of magazines to create, default: 1
disk-type
specifies the type of disks to create, default depends on the type of the cage
See the section Add a Cage
Examples:
1. add two magazines with default disk-type:
# esd add mag –w 50050cc2fc464dc4 mag-num=2
The above command adds two magazines to the cage with wwn 50050cc2fc464dc4. If the cage is DCN2 or DCS8, FC drives will be added. If
the cage is DCS7, NL drives will be added.
2. Add 1 magazine with SSD drive:
# esd add mag –w 50050cc2fc464dc4 disk-type=SSD
The above command adds 1 magazine with SSD drive to the cage with wwn 0x50050cc2fc464dc4. If the cage is DCS7, this operation will fail
because DCS7 doesn’t allow SSD drives.
Technical white paper
Page
14
Add Disk(s)
Synopsis:
esd add disk -w <cage-wwn> [disk-type={FC|SSD|NL}] [disk-num=<num>]
-w <cage-wwn> specifies which cage to add the disk
disk-numspecifies the number of disks to create, default: 1
disk-type
specifies the type of disks to create, default depends on the type of the cage
See the section Add a Cage
Examples:
1.
2.
Add 8 disks with default disk-type:
# esd add disk –w 50050cc2fc464dc4 disk-num=8
The above command adds 8 drives to the cage with the wwn 50050cc2fc464dc4.
If the cage is DCN2 or DCS8, FC drives will be added, if the cage is DCS7, NL drives will be added.
Add 1 NL drive:
# esd add disk –w 50050cc2fc464dc4 disk-type=NL
The above command adds 1 magazine with SSD drive to the cage with the wwn 50050cc2fc464dc4.
If the cage is DCS8, this operation will fail because DCS8 doesn’t allow NL drives.
Delete a Cage
Synopsis:
esd delete cage -w <cage-wwn>
Example:
# esd delete cage –w 50050cc2fc464dc4
Note
The above command deletes the cage with wwn 50050cc2fc464dc4, the corresponding data directory for this cage will be removed as well.
Delete a Disk
Synopsis:
esd delete disk –w <disk-wwn>
Example:
# esd delete disk –w 20000002fc464dc4
Note
The above command deletes a disk with wwn 20000002fc464dc4, if the magazine becomes empty, it will be deleted as well. The corresponding
data file for this disk will be removed.
List Cage(s)
Synopsis:
esd list cage [-w <cage-wwn>]
Note
If -w <cage-wwn> is specified, it lists detailed information of the specific cage; otherwise, it lists a summary of all cages.
Examples:
1) List summary of all cages, one line per cage:
# esd list cage
CageType
WWN
Drives Vendor
ProdID
DCN2
50050cc2fc47cff4 23
PTzk5fVt EB-2425P-E6EOS 3207
DCS7
50050cc2fc48ddc8 7
PTzk5fVt HB-2435-E6EBD
DCS8
50050cc2fc4ae71810
PTzk5fVt EB-2425P-E6EBD 3207
Rev
3207
Note
The list only displays all cages configured by esd, it is not necessarily ordered as “showcage” command on the cluster node.
Technical white paper
Page
15
2)
List the detailed information of a specific cage:
# esd list cage –w 50050cc2fc47cff4
List Magazine(s)
Synopsis:
esd list mag -w <cage-wwn>
Note
This lists all magazines of a specific cage. Since there is one disk per magazine in an EOS system, this lists all disks of a cage.
Example:
List all magazines (i.e. disks) of a cage:
# esd list mag –w 50050cc2fc47cff4
===== disk =====
ID
DEVNAME
Serial# NumOfBlks
0:0
20000002fc47cff4 bb000000 943718400
520
1:0
20000002fc47d0f4
bb000100 943718400
2:0
20000002fc47d1f4
bb000200 943718400 520
BlkSz
Vendor Model
SEAGATE ST9450404SS
XRFA
520
SEAGATE ST9450404SS
SEAGATE ST9450404SS
XRFA
Rev
ok
XRFA
ok
Status
20
ok
20
Temp
List a Disk
Synopsis:
esd list disk –w <disk-wwn>
Example:
List information of a specific disk:
# esd list disk –w 20000002fc47cff4
===== disk =====
ID
DEVNAME
Serial#
0:0
20000002fc47cff4 bb000000
BlkSz
520
Rev
XRFA
Status
ok
Temp
20
NumOfBlks
943718400
Vendor Model
SEAGATE ST9450404SS
20
ESD Log FIles
The esd log file location is /var/log/esd. It logs the following information in the esd.log* files:
•
Daemon start/stop time along with its pid
•
connections with vhba
•
errors
The esd keeps last 5 log files (from esd.log.0 to esd.log.4). The esd.log has the most recent information. If a change configuration fails, check esd.log
for errors.
Exporting Virtual Volumes
The HP 3PAR Simulator environment does not support a host entity. To export volumes to a host, a fake host entity must be created. The
simulator provides FC (Fibre Channel) ports for the fake host to connect and simulate host connectivity only, host I/O is not supported with this
connection.
Before exporting a virtual volume, you need to simulate the host connection.
1. Log in to a cluster node with user console and password crH72mkr
2. Choose option 3 (Simulator Host Connection), this will create a host WWN.
You can then use the 3PAR Management Console or Remote CLI to create host attributes and export the volume. If you use Remote CLI, run
showhost to find the host WWN. The host WWN must be used with the createhost command.
Note
1. The host connection simulation is an internal utility and may be changed/obsoleted in the future releases of HP 3PAR Simulator.
2. The simulated host connections once created they can not be deleted.
3. The simulator host connections are not persistent through simulator reboots. they are automatically removed upon rebooting the nodes
Technical white paper
Page
16
Refer to CLI Administrator's Manual for the full syntax of the commands createhost, createvlun to export volumes. See the section Error:
Reference source not found
Remote Copy
Remote Copy allows copying virtual volumes from one storage system to another.
With the HP 3PAR Simulator:
• Only Remote Copy over IP (RCIP) and a 1:1 configuration is supported
•
•
Any RC configuration that requires more than 2 nodes in a cluster is not supported
Assign one of the two supported serial numbers to each instance of the simulator. The serial number assignment is part of the cluster node
configuration & setup, ensure the same serial number is assigned to both the cluster nodes in the same instance of the simulator.
Requirements
To run a 1:1 RCIP configuration in a server or workstation environment, your system needs to be able to support:
•
A total of 6 VMs, three VMs for each simulated storage system.
•
The cluster node VMs require 2 vCPUs and 4GB memory each.
•
The two instances of the simulator can be run in a single system or two different systems. The RC network connectivity between the two
simulator instances must be setup via a separate vSwitch on ESXi or LAN segment on VMware Workstation 9. The vNIC#2 in each of the cluster
nodes must be connected to the vSwitch or the LAN segment.
The following requirements are general RC requirement, not specific to simulator:
•
one static cluster IP address for each cluster
•
One static IP address for each RC port, the RC IP address must be in a different subnet than the primary node and cluster IP address (vNic#2 in
each cluster node VM).
Configuration
Follow instructions in RC user guide to setup the RC over IP configuration. See the section Error: Reference source not found for reference to the
RC user guide.
As an example, the section below in the document is to illustrate how the RCIP can be setup using the HP 3PAR CLI. Users might choose differtent
IPs and subnet masks.
Important
In the following instructions, Cluster A refers to the 2-node cluster in Storage System 1 and Cluster B refers to the 2-node cluster in Storage
System 2
Cluster A:
First setup the netc configuration to configure cluster IP; this is required for the RC commands to work
netcconf -set
Follow the prompts to enter IP address, subnet mask and gateway address
To verify the configuration, use the command shownet
RC configuration:
controlport rcip addr 192.1.82.75 255.255.248.0 0:3:1
controlport rcip addr 192.1.82.76 255.255.248.0 1:3:1
controlport rcip gw 192.1.80.1 0:3:1
controlport rcip gw 192.1.80.1 1:3:1
Use showport –rcip to verify the RC port information
Technical white paper
Page
17
Cluster B:
First setup the netc configuration; this is required for the RC commands to work
netcconf –set
Follow the prompts to enter IP address, subnet mask and gateway address
To verify the configuration, use the command shownet
RC configuration:
controlport rcip addr 192.1.82.77 255.255.248.0 0:3:1
controlport rcip addr 192.1.82.78 255.255.248.0 1:3:1
controlport rcip gw 192.1.80.1 0:3:1
controlport rcip gw 192.1.80.1 1:3:1
Use showport –rcip to verify the RC port information
Cluster A:
To verify the RC link connectivity:
controlport rcip ping 192.1.82.77 0:3:1
controlport rcip ping 192.1.82.78 0:3:1
To start remote copy and create target:
startrcopy
creatercopytarget srcp2 IP 0:3:1:192.1.82.77 1:3:1:192.1.82.78
To create a CPG and a VV:
createcpg -ha mag -sdgs 16g -sdgl 32g -sdgw 24g cpg1
createvv -snp_cpg cpg1 cpg1 vv1 10g
Cluster B:
To verify the RC link connectivity:
controlport rcip ping 192.1.82.75 0:3:1
controlport rcip ping 192.1.82.76 0:3:1
To start remote copy and create target:
startrcopy
creatercopytarget srcp1 IP 0:3:1:192.1.82.75 1:3:1:192.1.82.76
To create a CPG and a VV:
createcpg -ha mag -sdgs 16g -sdgl 32g -sdgw 24g cpg1
createvv -snp_cpg cpg1 cpg1 rvv1 10g
Cluster A:
To create an rcopy group and admit rcopy VV:
creatercopygroup rc_group1 srcp2:periodic
admitrcopyvv vv1 rc_group1 srcp2:rvv1
Start initial replication:
startrcopygroup rc_group1
Use showrcopy to display the status of the replication
To resync:
syncrcopy rc_group1
Use showrcopy to check the status of the replication
The HP 3PAR Management Console can also be used to configure RCIP. Please refer to the Remote Copy and Managemement user guides for more
information.
Technical white paper
18
The HP 3PAR StoreServ Simulator sharepoint
http://ent142.sharepoint.hp.com/teams/3ParSimulator/SitePages/Home.aspx
Page
Technical white paper
19
Resources, contacts, or additional links
hp.com/vanity/url
Page
Download