HPE Storage Users Group
https://3parug.net/

3Par Fibre Channel Virtual Volume shows as ISCSI
https://3parug.net/viewtopic.php?f=18&t=428
Page 1 of 1

Author:  mwkirk [ Wed Sep 25, 2013 5:36 pm ]
Post subject:  3Par Fibre Channel Virtual Volume shows as ISCSI

We are having an issue with a 3Par host where we have a 12 TB VV presented as an RDM to a Windows 2012 Vmware VM.

This drive appears to have poor performance compared with other similar Drives and this same VM has another RDM that appears to perform normally.

The ONLY thing I see different is when we look at this LUN in devices in Vmware as a 3PARdata ISCSI Disk but has a valid FC identifier of naa.60002ac0000000000000000900003d6e. Also looking in some of the performance counters in Vmware this Lun shows high Physical Device Latency at times.

The disk showing up as ISCSI may be a red herring but at this point I am just looking at one of these things is not like the others.

Anyone ever seen anything like this?

Author:  hdtvguy [ Thu Sep 26, 2013 6:18 am ]
Post subject:  Re: 3Par Fibre Channel Virtual Volume shows as ISCSI

What SCSI controller are you using in the VM? Paravirtual SCSI would be the recommended controller. What are you using to gauge performance for the RDM other than vmware perf counters, have you run any Windows perfmon counters? Are the NTFS cluster sizes the same for both RDMs?

Author:  Richard Siemers [ Fri Sep 27, 2013 12:10 am ]
Post subject:  Re: 3Par Fibre Channel Virtual Volume shows as ISCSI

Can you confirm the lun settings in the 3par side to check into the FC vs ISCSI export?

Author:  mwkirk [ Wed Oct 30, 2013 7:24 am ]
Post subject:  Re: 3Par Fibre Channel Virtual Volume shows as ISCSI

It was presented as Fibre-Channel but had been previously presented as ISCSI so according to HP that is why it picked up the ISCSI disk identifier and it should not really affect anything and is just a name.

Author:  Architect [ Tue Nov 05, 2013 4:43 pm ]
Post subject:  Re: 3Par Fibre Channel Virtual Volume shows as ISCSI

maybe slightly offtopic, but If you spot less then optimal performance issues, please look at the growth sizes and step sizes of your CPGs.
- growth sizes for FC and NL should be 32GB*the number of nodepairs (so assuming a 4-node 7400, it should be 64GB). For SSD it should be 8GB*the number of nodepairs (16GB for a 4-node 3PAR)
- default step sizes depend on Raid level used, the set-size and drive type. Often I advice to remove the -ss value all together from the CPG configuration, to force the 3PAR to use the defaults (which are always right). Of course there are very heavily tuned CPGs possible where you could need to vary the stepsize based on the underlying application, but usually the default is best.

e.g. via CLI do the following to show the CPG configuration:
Code:
showcpg -sdg

output would be something like:
Code:
Id Name                  Domain            Warn   Limit  Grow Args
Some_FC_Raid5_CPG        SomeDomain           -       - 16384 -t r5 -ha cage -ssz 4 -ss 32 -ch first -p -devtype FC 


in this example both the growth size is to small 16GB instead of 64GB), and the step size is wrong (32KB vs 128KB). setting the right value is done like this:

Code:
setcpg -sdgs 65536 -t r5 -ha cage -ssz 4 -ss 128 -ch first -p -devtype FC Some_FC_Raid5_CPG
 


The system will require a "tunesys" afterwards to adjust the right stepsizes for all previously created LDs with the "wrong" specifications. Please suspend AO during a tunesys to prevent nasty issues.

Below a reference with the right stepsizes per drivetype / raid / setsize:

For HDDs (both NL and FC):
Code:
R1:     256k
R5:     128k
R6 4+2: 128k
R6 6+2:  64k
R6 8+2:  64k
R6 10+2: 64k
R6 14+2: 32k

For SSD's:
Code:
R0:      32k
R1:      32k
R5:      64k
R6 6+2:  64k
R6 8+2:  64k
R6 10+2: 64k
R6 14+2: 32k

Page 1 of 1 All times are UTC - 5 hours
Powered by phpBB © 2000, 2002, 2005, 2007 phpBB Group
http://www.phpbb.com/