Ibm V7000 Multipath Drivers For Mac

  1. Ibm V7000 Unified
  2. Epson Drivers For Mac
  3. Ibm V7000 Datasheet
  • Cluster and VMware Solution Implementation Angelo Bernasconi Torben Jensen Ian MacQuarrie Ole Rasmussen. Iv IBM SAN and SVC Stretched Cluster and VMware Solution Implementation. (PSP) and Native MultiPath Drivers (NMP)..........
  • To ibm v7000 multipath more information about the storage systems controlled by IBM Flex System Manager and see which actions there can be performed, click Network Storage in the storage control window, as shown in Figure Take Offline in muotipath Actions drop-down menu.
  • Jul 17, 2013  Hi - we have installed an IBM v7000 SAN using 4 X 1Gbps iSCSI connections, 2 from each controller card. Multipath is working as I.


The iSCSI storage driver for IBM Storwize V7000 SAN provides OpenNebula with the possibility of using V7000 volumes as block devices for VM images. The use of block based storage presents several benefits over image based storage, especially regarding performance. The entire SAN volume management is achieved through the OpenNebula front-end. This driver supports multipathing, replication and failover with two SAN boxes. Only an iSCSI client and multipath setup is required on host side.



This add-on is compatible with OpenNebula 4.6+


An optional deployment host is recommended to prepare virtual disks (mkfs, image upload...) but this can be handled by the front-end as well. The benefits are you can take this workload off the front-end and you only have to restart deployment host in case of trouble, not the whole front-end. The deployment host can of course be an OpenNebula VM. You can also decide to use multiple deployment hosts. They will be selected by round-robin like algorithm. If you decide not to use the deployment host(s), its prerequisites apply to the front-end.

Ibm V7000 Multipath Drivers For Mac

We have long experience running both options but currently we are moving away from rdac mpp multipath driver and relying only for Linux native dmp multipath, eg. IBM Storwize V7000 series needs dmp multipath.

OpenNebula Front-end

Ibm V7000 Unified

  • ssh with public key authentication to IBM V7000 for oneadmin user
  • flock which is part of util-linux package

flock is used to manage v7000 access concurrency. Only one session will be opened on a single v7000 box at a time.

OpenNebula Deployment Host(s)

  • open-iscsi initiator
  • multipath
  • ddpt

ddpt is used to clone to a different target datastore using sparse copy to speed up data transfers. You can still use traditional dd command if you prefer. See the beginning of file and set USE_DDPT=0.

OpenNebula Hosts

  • open-iscsi initiator
  • multipath

IBM Storwize V7000

  • A oneadmin user with public key authentication
  • Registered OpenNebula hosts and deployment host(s) for iSCSI.


OpenNebula Front-End

  • Copy these files/directories:

    • datastore/v7000 -> /var/lib/one/remotes/datastore/v7000
    • tm/v7000 -> /var/lib/one/remotes/tm/v7000
  • Add 'v7000' to 'arguments' attribute of TM_MAD and DATASTORE_MAD in /etc/one/oned.conf

  • Add the following to the bottom of /etc/one/oned.conf

  • For live migration, you have to add the following to

    • /var/lib/one/remotes/tm/shared/premigrate

    • /var/lib/one/remotes/tm/shared/postmigrate


###Configuring the System Datastore


To use V7000 drivers, you have to configure the system datastore as shared. This sytem datastore will only hold the symbolic links to the block devices, so it will not take much space. See more details on the System Datastore Guide.

It will also be used to hold context images, they will be created as regular files.

Configuring iSCSI Datastores

The first step to create an iSCSI datastore is to set up a template file for it. In the following table you can see the supported configuration attributes. The datastore type is set by its drivers, in this case be sure to add DS_MAD=v7000 and TM_MAD=v7000 for the transfer mechanism, see below. The options regarding V7000 volumes management are detailed in the IBM SVC and Storwize V7000 V640 CLI Guide.

Mandatory configuration attributes

  • NAME: [name] The name of the datastore.
  • DS_MAD: [v7000] The DS type, use v7000 for the V7000 datastore.
  • TM_MAD: [v7000] Transfer driver for the datastore, use v7000, see below
  • BRIDGE_LIST: [fqdn] The deployment host(s) FQDN. Defaults to localhost.localdomain.
  • MGMT : [fqdn] The V7000 master box FQDN. The default value is v7000-master.localdomain.

Optionnal configuration attributes

  • BASE_IQN[iqn] The base IQN for V7000 iSCSI target. The default value is
  • NO_ISCSI[hostname] Lists of hosts (separated by spaces) for which no iscsiadm login or logout is performed. Defaults to $HOSTNAME.
  • THIN_PROVISION[1 0] Activate thin-provisioned volumes. The default value is 1.
  • SNAPSHOT[1 0] Use snapshots or clone for non-persistent images. The default value is 1.
  • SNAPSHOT_RSIZE[%](mkvdisk -rsize option) Use with SNAPSHOT=1 parameter. Defines how much physical space is initially allocated to the thin-provisioned volume for non persistent image. The default value is 0 %.
  • RSIZE[%](mkvdisk -rsize option) Defines how much physical space is initially allocated to the thin-provisioned volumes for persistent images. The default value is 2 %.
  • COPIES[1 2](mkvdisk -copies option) Specifies the number of local volume copies to create. Setting the value to 2 creates a mirrored volume. The default value is 1.
  • IO_GROUP[io_grp](mkvdisk -iogrp option) Specifies the I/O group (node pair) with which to associate volumes. The default value is io_grp0.
  • MDISK_GROUP[mdisk_grp](mkvdisk -mdiskgrp option) Specifies one or more managed disk groups (storage pools) to use when creating volumes. The default value is mdiskgrp0.
  • SYNC_RATE[0 - 100](mkvdisk -syncrate option) Specifies the copy synchronization rate for volumes. A value of zero (0) prevents synchronization. The default value is 50.
  • VTYPE[striped](mkvdisk -vtype option) Specifies the virtualization type. The default virtualization type is striped and it is the only type currently implemented in this driver.
  • COPYRATE[0 - 100](mkfcmap -copyrate option) Specifies the copy rate for the clone mapping. The rate value can be 0 - 100. A value of 0 indicates no background copy process. The default value is 50
  • CLEANRATE[0 - 100](mkfcmap -cleanrate option) Sets the cleaning rate for the clone mapping. The rate value can be 0 - 100. The default value is 50.
  • GRAINSIZE[64 256](mkfcmap -grainsize option) Specifies the grain size for the clone mapping. The default value is 256. Once set, this value cannot be changed. The default value is 256

Configuration attributes for replication and failover with auxiliary box setup

  • MGMT_AUX[fqdn] The V7000 auxiliary box FQDN for replication. The default value is v7000-aux.localdomain.
  • CLUSTER[cluster_id] The V7000 Cluster ID. The default value is 00000200A0000000.
  • REPLICATION[1 0] Used to activate replication one the slave box. The default value is 0.
  • FAILOVER[1 0]Use auxiliary volumes instead of master ones on the auxiliary box. Volumes one the auxiliary box become masters. The default value is 0.

Configuring v7000 driver options

You can tweak some v7000 driver options inside file. The comments are self-explanatory.

Configuring OpenNebula Hosts and Deployment Host(s)

Firstly you need to know/configure your initiator name, see /etc/iscsi/initiatorname.iscsi. Recommended open-iscsi settings in /etc/iscsi/iscsid.conf are:

Secondly you have to configure multipath daemon. Here is the configuration recommended by IBM for /etc/multipath.conf. It is also recommended to blacklist local disks. Here, it is /dev/sda.

Finally, configure the following permissions for oneadmin user in /etc/sudoers:

You have to restart open-iscsi and multipath services afterwards.

All is left to do is to register and login to all V7000 nodes with

iscsiadm -m discovery -t st -p $NODE_IP:3260 --op update -n node.startup -v automatic

Where $NODE_IP is the ip of one of your nodes. So you wil have to execute the command for each $NODE_IP. The settings will be retained and the iSCSI sessions automatically restarted at boot.

Configuring IBM Storwize V7000

  • Create a oneadmin user with public key authentication
  • Register IQN of OpenNebula hosts and deployment host(s)


I will illustrate the most complete solution: the dual box setup. If you want to achieve active-active setup, you can configure two datastores, one for each box. For each datastore you configure the other box as auxiliary.



Whith REPLICATION=1 a replicated auxiliary volume will be created on MGMT_AUX. This volume is read-only, this is inherent to V7000 design. If you loose the master box, you will loose the VMs whose disks are mapped from this box. You can quickly recover by restarting the VMs and make them use the auxiliary volumes simply by setting FAILOVER=1 datastore option.

Epson Drivers For Mac

Once you have recovered from the incident, you can set back FAILOVER=0. All running VMs will continue to use auxiliary volumes but they will switch to master volumes on next deployment. If you decide to create new volumes while FAILOVER=1, it is your responsibility to manage replication as soon as the other box is back online.


You can configure jumbo frames (MTU=9000) on both V7000 and iSCSI initiators. Please, make sure jumbo frames are supported by your network equipment (switches, routers...)You can also achieve a big performance improvement by using noop I/O scheduler inside VMs and deadline I/0 scheduler on the hosts instead of cfq. Configure it with elevator=noop or elevator=deadline kernel options.

Hiya Everyone,
Just looking for some advice. Its been an issue for a while, and although i have a workaround, i want to know the opinion of someone who knows a little more about this.
I have some HS22 Blades, in a BladeCenter E Chasis, which connect to a V7000 IBM SAN.
My problem is, i have 2 Fibre cards in each Blade. These have two viewable mac addresses in the Storage Explorer. If i map both MAC's to a host, the volume will show up twice. I have the Multipath I/O installed and the HBA drivers setup.
Im not sure if this is a failover feature im looking for outside of the functionality of this system, or if i am doing something wrong.
We had a technical consultant come in, and only map 1 MAC address to the Host, meaning it was showing once, but multipath doesnt appear to work. Because if you pull out a fibre connection, the systems go offline.
Any advice on this would be very much appreciated.

Ibm V7000 Datasheet

Windows Server 2008 R2 SP1
Comments are closed.