Quantcast
Channel: MSA Storage topics
Viewing all 2299 articles
Browse latest View live

MSA1500 - Active/Active Yet?

$
0
0
Hi Guys,

Just after confirmation - my understanding is that MSA1500 with the current FW V.4.98 does not support Active/Active controller configuration. Could someone please confirm this for me. I obviously do not get to play much with the MSA1000/1500s :(

Anyone knows when V5.x of the FW is expected and will it definitely bring the controllers to behave "Active/Active" for different LUNs?

Regards,
Saket.

symmpi errors on Windows 2003 using MSA-1000 LUNS

$
0
0
Hi There,

I'm trying to troubleshoot some issues with a Windows 2003 server that's using MSA-1000 storage. It's an exchange 2003 server that clients reporting freezing for seconds a few times a day. The only suspect event log entries on the Windows side is:

15 The device, \Device\Scsi\symmpi1, is not ready for access yet.

11 The driver detected a controller error on \Device\Harddisk1.


These occur about 30 times in a row, twice a day. I've correlated this with backup times etc, the only clues I have is 9.00 is about peak time and 4.30 is around the time our exchange backups kick off.

I'm wondering how to troubleshoot this problem?

Adding new MSA1000 to existing RA4100 San

$
0
0
Hi,
I need to add a new MSA shelf to our exsting san off which a couple of sql clusters run. The plan I have for doing this is as follows :
1. Connect MSA up to both san switches.
2. Power MSA on.
3. Identify sql node 1 to storage through SSP.
4. Scan for new hardware on sql node 1 and install MSA.
5. Create raid sets and partitions on storage from node 1.
6. Create new cluster disk group with preferred owner as node 1.
7. Identify sql node 2 to MSA via SSP.
8. Scan for new hardware and install MSA on node 2.
9. If not automatically done, create redundant paths via secure path manager on node 1.

Just wondering if this is the correct order to do things, whteher anything is missing and whether i'd actually have to power off sql node 2 before adding the storage.
Any help is much appreciated !!
Thanks, Dave.

MSA2040 VDISK SPARE

$
0
0

We have a couple of hundred of these enclosures around the country some with a D2700. Our standard configuration where we also have a D2700 attached, the vdisks are RAID6 and there's a global spare assigned to each enclosure. I don't know when it happened but we have one MSA with a D2700 that has a vdisk spare assigned to vd03, one global spare on the 2040, and no global spare on the D2700. This is not our standard configureration. We're unsure what would happen to the integrety of the MSA and the vdxx volums if we were to backoff this the vdisk spare on vd03 and make it a global spare on the D2700.  

HP MSA 2040 SAN crashing

$
0
0

I have an MSA 2040 SAN with dual controller and a second enclosure.  They have worked fine since I installed them in late 2014.  Tuesday, the SAN crashed, and the primary enclosure showed a fault light.  Trying to access the web interface said "Initialization failed".  I rebooted the system, and it came up normal and ran fine.  That is until today.  Same exact thing happened.  I have made no hardware or software changes in the last few weeks.

Below are my error logs from Enclosure A.  I can't really decipher what is going on.  It appears one of my servers is constantly trying to log into the SAN and failing.  I can't think of why it would be, and not sure if it is related.  I am only attaching up to the end of day on Oct 3, because today's logs are basically EXACTLY the same.

FIRMWARE:

Bundle Version: GL105P001-03

Storage Controller: GLS105R04-01

Storage Controller Loader: 27.015

Management Controller: GLM105R009-01

Management Controller Loader: 6.04.16617

Expander: 3117

CPLD: 42

Sep 28 11:22:36 (none) auth.info sshd[10081]: Failed password for root from 192.168.30.91 port 12452 ssh2

Sep 28 11:22:39 (none) auth.err sshd[10081]: error: PAM: Permission denied for root from 192.168.30.91

Sep 28 11:22:39 (none) auth.info sshd[10081]: Connection closed by 192.168.30.91 [preauth]

Sep 28 15:16:49 (none) auth.info sshd[14885]: Did not receive identification string from 192.168.30.91

Sep 28 15:27:50 (none) auth.info sshd[15114]: Did not receive identification string from 192.168.30.91

Sep 28 15:27:53 (none) auth.info sshd[15117]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Sep 28 15:27:56 (none) auth.info sshd[15117]: Failed password for root from 192.168.30.91 port 24622 ssh2

Sep 28 15:27:58 (none) auth.err sshd[15117]: error: PAM: Permission denied for root from 192.168.30.91

Sep 28 15:27:58 (none) auth.info sshd[15117]: Connection closed by 192.168.30.91 [preauth]

Sep 28 19:16:09 (none) auth.info sshd[19801]: Did not receive identification string from 192.168.30.91

Sep 28 23:16:17 (none) auth.info sshd[24725]: Did not receive identification string from 192.168.30.91

Sep 29 03:16:49 (none) auth.info sshd[29637]: Did not receive identification string from 192.168.30.91

Sep 29 07:16:54 (none) auth.info sshd[2104]: Did not receive identification string from 192.168.30.91

Sep 29 07:22:22 (none) auth.info sshd[2260]: Did not receive identification string from 192.168.30.91

Sep 29 07:22:26 (none) auth.info sshd[2263]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Sep 29 07:22:30 (none) auth.info sshd[2263]: Failed password for root from 192.168.30.91 port 14036 ssh2

Sep 29 07:22:34 (none) auth.err sshd[2263]: error: PAM: Permission denied for root from 192.168.30.91

Sep 29 07:22:34 (none) auth.info sshd[2263]: Connection closed by 192.168.30.91 [preauth]

Sep 29 11:16:50 (none) auth.info sshd[7142]: Did not receive identification string from 192.168.30.91

Sep 29 15:16:54 (none) auth.info sshd[12035]: Did not receive identification string from 192.168.30.91

Sep 29 19:16:07 (none) auth.info sshd[16928]: Did not receive identification string from 192.168.30.91

Sep 29 23:16:23 (none) auth.info sshd[21845]: Did not receive identification string from 192.168.30.91

Sep 30 03:16:49 (none) auth.info sshd[26769]: Did not receive identification string from 192.168.30.91

Sep 30 03:22:06 (none) auth.info sshd[26883]: Did not receive identification string from 192.168.30.91

Sep 30 03:22:10 (none) auth.info sshd[26886]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Sep 30 03:22:13 (none) auth.info sshd[26886]: Failed password for root from 192.168.30.91 port 15958 ssh2

Sep 30 03:22:17 (none) auth.err sshd[26886]: error: PAM: Permission denied for root from 192.168.30.91

Sep 30 03:22:17 (none) auth.info sshd[26886]: Connection closed by 192.168.30.91 [preauth]

Sep 30 07:15:32 (none) auth.info sshd[31672]: Did not receive identification string from 192.168.30.91

Sep 30 11:16:47 (none) auth.info sshd[4305]: Did not receive identification string from 192.168.30.91

Sep 30 11:22:00 (none) auth.info sshd[4410]: Did not receive identification string from 192.168.30.91

Sep 30 11:22:04 (none) auth.info sshd[4418]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Sep 30 11:22:07 (none) auth.info sshd[4418]: Failed password for root from 192.168.30.91 port 30993 ssh2

Sep 30 11:22:10 (none) auth.err sshd[4418]: error: PAM: Permission denied for root from 192.168.30.91

Sep 30 11:22:10 (none) auth.info sshd[4418]: Connection closed by 192.168.30.91 [preauth]

Sep 30 15:16:36 (none) auth.info sshd[9230]: Did not receive identification string from 192.168.30.91

Sep 30 19:16:10 (none) auth.info sshd[14139]: Did not receive identification string from 192.168.30.91

Sep 30 23:16:27 (none) auth.info sshd[19071]: Did not receive identification string from 192.168.30.91

Oct  1 03:16:44 (none) auth.info sshd[23985]: Did not receive identification string from 192.168.30.91

Oct  1 07:15:43 (none) auth.info sshd[28878]: Did not receive identification string from 192.168.30.91

Oct  1 11:16:56 (none) auth.info sshd[1336]: Did not receive identification string from 192.168.30.91

Oct  1 15:16:45 (none) auth.info sshd[6384]: Did not receive identification string from 192.168.30.91

Oct  1 19:16:07 (none) auth.info sshd[11289]: Did not receive identification string from 192.168.30.91

Oct  1 23:15:18 (none) auth.info sshd[16192]: Did not receive identification string from 192.168.30.91

Oct  1 23:19:21 (none) auth.info sshd[16279]: Did not receive identification string from 192.168.30.91

Oct  1 23:19:24 (none) auth.info sshd[16282]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Oct  1 23:19:28 (none) auth.info sshd[16282]: Failed password for root from 192.168.30.91 port 60828 ssh2

Oct  1 23:19:31 (none) auth.err sshd[16282]: error: PAM: Permission denied for root from 192.168.30.91

Oct  1 23:19:31 (none) auth.info sshd[16282]: Connection closed by 192.168.30.91 [preauth]

Oct  2 03:16:47 (none) auth.info sshd[21152]: Did not receive identification string from 192.168.30.91

Oct  2 07:15:43 (none) auth.info sshd[26046]: Did not receive identification string from 192.168.30.91

Oct  2 11:16:52 (none) auth.info sshd[30993]: Did not receive identification string from 192.168.30.91

Oct  2 15:16:14 (none) auth.info sshd[3574]: Did not receive identification string from 192.168.30.91

Oct  2 15:20:26 (none) auth.info sshd[3660]: Did not receive identification string from 192.168.30.91

Oct  2 15:20:30 (none) auth.info sshd[3663]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Oct  2 15:20:34 (none) auth.info sshd[3663]: Failed password for root from 192.168.30.91 port 15754 ssh2

Oct  2 15:20:37 (none) auth.err sshd[3663]: error: PAM: Permission denied for root from 192.168.30.91

Oct  2 15:20:37 (none) auth.info sshd[3663]: Connection closed by 192.168.30.91 [preauth]

Oct  2 19:16:09 (none) auth.info sshd[8502]: Did not receive identification string from 192.168.30.91

Oct  2 23:16:11 (none) auth.info sshd[13431]: Did not receive identification string from 192.168.30.91

Oct  3 03:16:20 (none) auth.info sshd[18345]: Did not receive identification string from 192.168.30.91

Oct  3 07:15:14 (none) auth.info sshd[23234]: Did not receive identification string from 192.168.30.91

Oct  3 11:16:21 (none) auth.info sshd[28174]: Did not receive identification string from 192.168.30.91

Oct  3 11:20:37 (none) auth.info sshd[28266]: Did not receive identification string from 192.168.30.91

Oct  3 11:20:41 (none) auth.info sshd[28269]: WARNING: /home/pablo/sandbox/heartbleed/GL105/platform/build.Gallium.GPM105P01-01.output/opt/etc/moduli does not exist, using fixed modulus

Oct  3 11:20:44 (none) auth.info sshd[28269]: Failed password for root from 192.168.30.91 port 46661 ssh2

Oct  3 11:20:48 (none) auth.err sshd[28269]: error: PAM: Permission denied for root from 192.168.30.91

Oct  3 11:20:48 (none) auth.info sshd[28269]: Connection closed by 192.168.30.91 [preauth]

Oct  3 12:05:00 (none) user.warn kernel: sched: RT throttling activated

Oct  3 12:05:00 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:21 (none) user.warn kernel: ------------[ cut here ]------------

Oct  3 12:05:21 (none) user.warn kernel: WARNING: at net/sched/sch_generic.c:256 dev_watchdog+0x15c/0x24c()

Oct  3 12:05:21 (none) user.info kernel: NETDEV WATCHDOG: mcmc (): transmit queue 0 timed out

Oct  3 12:05:21 (none) user.warn kernel: Modules linked in: mcmclink g_serial ocores_udc mcfulink mooseproc mcscbridge msgdrv

Oct  3 12:05:21 (none) user.warn kernel: [<c0014428>] (unwind_backtrace+0x0/0xec) from [<c02a6c50>] (dump_stack+0x20/0x24)

Oct  3 12:05:21 (none) user.warn kernel: [<c02a6c50>] (dump_stack+0x20/0x24) from [<c001bf60>] (warn_slowpath_common+0x5c/0x74)

Oct  3 12:05:21 (none) user.warn kernel: [<c001bf60>] (warn_slowpath_common+0x5c/0x74) from [<c001c034>] (warn_slowpath_fmt+0x40/0x48)

Oct  3 12:05:21 (none) user.warn kernel: [<c001c034>] (warn_slowpath_fmt+0x40/0x48) from [<c022a1e4>] (dev_watchdog+0x15c/0x24c)

Oct  3 12:05:21 (none) user.warn kernel: [<c022a1e4>] (dev_watchdog+0x15c/0x24c) from [<c00281b8>] (run_timer_softirq+0x1d0/0x2dc)

Oct  3 12:05:21 (none) user.warn kernel: [<c00281b8>] (run_timer_softirq+0x1d0/0x2dc) from [<c0021cfc>] (__do_softirq+0xd8/0x1c0)

Oct  3 12:05:21 (none) user.warn kernel: [<c0021cfc>] (__do_softirq+0xd8/0x1c0) from [<c002219c>] (irq_exit+0x50/0x5c)

Oct  3 12:05:21 (none) user.warn kernel: [<c002219c>] (irq_exit+0x50/0x5c) from [<c000f750>] (handle_IRQ+0x84/0xa4)

Oct  3 12:05:21 (none) user.warn kernel: [<c000f750>] (handle_IRQ+0x84/0xa4) from [<c00086b8>] (asm_do_IRQ+0x18/0x1c)

Oct  3 12:05:21 (none) user.warn kernel: [<c00086b8>] (asm_do_IRQ+0x18/0x1c) from [<c000e55c>] (__irq_usr+0x3c/0x80)

Oct  3 12:05:21 (none) user.warn kernel: Exception stack(0xd5e5bfb0 to 0xd5e5bff8)

Oct  3 12:05:21 (none) user.warn kernel: bfa0:                                     001cc5f8 80000163 00000008 4390de30

Oct  3 12:05:21 (none) user.warn kernel: bfc0: 43aa46c0 00271970 41a2e8b0 b60fe65c 3bf37a15 43a7d83c 00000000 b60fe09c

Oct  3 12:05:21 (none) user.warn kernel: bfe0: 43946d98 b60fe088 41e66154 4390de30 20000010 ffffffff

Oct  3 12:05:21 (none) user.warn kernel: ---[ end trace 3867a107f846c95a ]---

Oct  3 12:05:21 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:21 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:21 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:30 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:31 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:33 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:39 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:40 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:53 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:55 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:05:59 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:00 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:04 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:05 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:06 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:11 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:13 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:17 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:19 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:23 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:24 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:29 (none) user.warn kernel: MCMC: error status (0xda) - Unexpected failure. (Inter MC link message(0x17))

Oct  3 12:06:42 (none) user.warn kernel: MCMC: error status (0x5f) - Message timed out. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0x5f) - Message timed out. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:53 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:54 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:06:59 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

Oct  3 12:07:00 (none) user.warn kernel: MCMC: error status (0xdc) - Memory write failed. (Inter MC link message(0x17))

SAN temperature

$
0
0
I have an ongoing problem with an MSA1000, the event viewer of the server (Windows 2003) connected to the SAN is filling up with errors, the error is as follows:

Sensor 1 of External Storage System on Port 0, Box 1 of the Array Controller \Device\FibreArray0, HBA Slot 6, Chassis: 0000JN7200 is reporting the internal temperature has exceed the preset limit. The controller may attempt to shut down power to the storage device and/or spin down the installed disk drives.

Almost every piece of hardware on the SAN has been replaced, the firmware has been upgraded and the server has the latest HBA drivers. Has anyone else had this issue, if so, how did you over come it. The SAN is directly connected to the server using the I/O module, bangalore have since told me this is not supported, but then they have told me alot of other stuff and am not sure whether to believe them (a collegue has an ongoing call regarding RDP with bangalore, and after a month have told him that dynamic disks are not supported)

Thanks.

MSA1500cs reporting "Error occurred reading RIS copy from ... (all drives)"

$
0
0
Hi,

We have a pretty new MSA1500cs. It's got two fibre I/O modules and it's running Firmware Version 7 (active/active).

There's one MSA 30 attached, fully loaded with 14x 300 GByte disks.

I've defined two Logical Drives, with one physical disk acting as the hot-spare for both drives.

HP's System Management Homepage reports that one drive has failed (I /believe/ it's the hot spare, but it's proving difficult to identify). There are no fault lights on the exterior of the MSA or on any of the disks.

If I use "hpacucli" to display some information, here's what I get (apologies for the length of this output):

===
=> ctrl all show

MSA1500 CS at atgcls-msa (sn: PB9840TX3TY01G, csn: SGA06520EY, wwn:
500508B30093F9A1)
Smart Array P400 in Slot 1 (sn: PA2240J9SU71KW)

=> ctrl sn=PB9840TX3TY01G pd all show

MSA1500 CS at atgcls-msa

array A (Failed)

physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, Failed, spare)

array B (Failed)

physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, OK)
physicaldrive 0:0 (box 0:bay 0, Parallel SCSI, 300 GB, Failed, spare)

=> ctrl sn=PB9840TX3TY01G pd all show detail

MSA1500 CS at atgcls-msa

array A (Failed)

physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AVKA0707
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: Failed
Drive Type: Spare Drive
Interface Type: Parallel SCSI
Transfer Mode: Async Wide
Size: 300 GB
Transfer Speed: 10 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AUJN0706
Model: COMPAQ BD30089BBA


array B (Failed)

physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: OK
Drive Type: Data Drive
Interface Type: Parallel SCSI
Transfer Mode: Ultra 3 Wide
Size: 300 GB
Transfer Speed: 160 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720ATE30706
Model: COMPAQ BD30089BBA
physicaldrive 0:0
Box: 0
Bay: 0
Status: Failed
Drive Type: Spare Drive
Interface Type: Parallel SCSI
Transfer Mode: Async Wide
Size: 300 GB
Transfer Speed: 10 MB/Sec
Rotational Speed: 10000
Firmware Revision: HPB1
Serial Number: DA01P720AUJN0706
Model: COMPAQ BD30089BBA

===

Notice how all the Bay IDs etc, and most of the disk serial numbers, are identical? This seems a little odd to me.

I've also attached an ADU report (I've omitted information about the internal Smart Array controller). Notice how all disks are reporting "Error occurred reading RIS copy...."?

This can't be good, can it? I've logged a call with HP's hardware support, but they've yet to come back with anything sounding like a diagnosis.

Has anyone got any ideas? I've asked for the serial cable to be plugged in, so I might be able to run some commands via that. BTW, this is attached directly to a RHEL box, so I can't use any Windows utilities.

FWIW, the Logical Drives I've created /appear/ to be working fine.

MSA1500 A/A

$
0
0
We have the following topology

Two Clusters (each with 2 nodes)and 2 standalone servers are connected to MSA1500
Each server has 3 fiber cards.Zoning is configured so that the first card of each server is connected to the first controller through the first switch and the second card to the second controller through the second switch.Third card is for backup purposes and is connected to another switch zoned properly with a tapelibrary.
Recenty we migrated from Active/Passive to Active/Active firmware. However now we are facing some serious performance issues mainly with the backup. (The SQL users still continue to complain as before).I have some questions.

1)Do we need to keep zoning in the fiber switches now with A/A firmware?
2)In the ACU, setting the preferred logical paths to each controller will cause temporal failover to the clusters?
3)We have installed SanSurfer. Should we change the rate settings of the fibercards from Auto to 2GB.Is it going to help? Are there any other parameters we could change there to increase performance.

4)Yesterday we run HPReadData and the results we got were frustrating
Server1 D: 3.69MB/Sec

Server1 Ε: 1,59ΜΒ/Sec

Server2 D: 9.76MB/Sec

Clusternode2 H: 0.43MB/Sec


5) The problem seems to be more intense with the cluster nodes running SQL.Now we need almost double time to backup sql databases.

Any suggestions or ideas would be very helpful.



MSA1000 Disk performance

$
0
0
Is there anyway to monitor the performance of the disk in an MSA1000 (ideally real time). We are experiencing queued back logs with our email environment and are trying to isolate where the bottleneck is (disk, fabric, communications,etc) The only thing I see is the ADU utility, But doesn't that really look at hardware failure more so then performance ? Insight manager reports no issues (both server and disk array)
Thanks;

Performance using MSA1500cs with MSA20 in ADG configuration.

$
0
0
Hi

What is the minimum and maximum write and read performance I should get from a raid 6 configuration?

In our configuration we get about 49Mb/s peak performance and about 30Mb/s in average.

Show tech_support output is attached.

The server is connected via fibre channel, and the fibre channel switch is only used for this server.

Server information:

Model: ProLiant DL380 G4
Serial numer: GB86111KLN
OS: Redhat / Red Hat Enterprise Linux AS release 4 (Nahant Update 5) / #1 SMP Fri Apr 20 17:03:35 EDT 2007
Fibre Channel: QLogic Corp. ISP2312-based 2Gb Fibre Channel to PCI-X HBA

Units are created using:

add unit 0 raid_level=ADG stripe_size=64 size=1572GB data="Disk101-Disk111"
add unit 1 raid_level=ADG stripe_size=64 size=1572GB data="Disk101-Disk111"
add unit 2 raid_level=ADG stripe_size=64 size=1572GB data="Disk101-Disk111"
add unit 3 raid_level=ADG stripe_size=64 size=1609159MB data="Disk101-Disk111"
add spare unit=0 "Disk112"

Filesystems are created with:

mkfs.ext3 -b 4096 -T largefile4 -m 0 -O sparse_super -O dir_index -R stride=16 -j -J size=400 /dev/sdm1
tune2fs -i 0 -c 0 /dev/sdm1
mkfs.ext3 -b 4096 -T largefile4 -m 0 -O sparse_super -O dir_index -R stride=16 -j -J size=400 /dev/sdn1
tune2fs -i 0 -c 0 /dev/sdn1
mkfs.ext3 -b 4096 -T largefile4 -m 0 -O sparse_super -O dir_index -R stride=16 -j -J size=400 /dev/sdo1
tune2fs -i 0 -c 0 /dev/sdo1
mkfs.ext3 -b 4096 -T largefile4 -m 0 -O sparse_super -O dir_index -R stride=16 -j -J size=400 /dev/sdp1
tune2fs -i 0 -c 0 /dev/sdp1


We performed the benchmark using dd, with direct file write to prevent write cache in Linux.

# cat /home/bigfile00 > /dev/null; sync; /bin/nice -n -10 /home/dd if=/home/bigfile00 of=/fc/unit200/fot/bigfile02 oflag=direct bs=1024k; sync
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 24.4965 seconds, 43.8 MB/s

MSA1500 SCSI redundant loop question

$
0
0
I have a MSA1500 and can’t find what I am looking for in the documentation. I want to run two loops to my disk shelfs for redundancy. I can’t find anything that specifically tells me that this is supported. Can anyone tell me if they have it setup this way?

Example:
MSA1500
4 SCSI I/O modules
2 SCSI shelf’s

Or
MSA1500
4 SCSI I/O modules
2 SATA shelf’s

connecting SAN msa 1500 to Blade 460c

$
0
0
I bought a Blade server bl460c and I want to connect ot my MSA1500. I don't know why I don't see my blade server when I run the HP arry utility on another server already connected to the SAN. I'm not the person who connected the first server to the MSA1500 so I only want to know if it is possible to have a little cook book.

Thank you

IBM AIX + Windows hosts and MSA 1000 (1500cs)

$
0
0
Dear All,

Is it possible to connect few Windows hosts and one IBM AIX host to MSA 1000 or MSA 1500cs
storage?

or EVA Familly and XP Family are only possible solutions.

Best regards
Goran

MSA1500

$
0
0
I have sporadically this error on the SAN!! What could this be?

MSA1500 Active/Active Successful

$
0
0
Hi.I have recently updated my MSA1500 Box to Active/Active. I just want to know from the professionals of this form that I did everything successfully or still I need some configurations in MSA1500 to boost up the performance.
Please have a look at show tech_support of both controllers in attached files.
Let me know what should I do more to get excellent performance from this Active/Active MSA1500?

MSA P2000 G3 FC migrate RAID-10 to RAID-6

$
0
0

Hi,

I would like to migrate a RAID-10 VDISK to a RAID-6 ! How can i perfom this please ?

I initialy thought of making a new RAID-6 VDISK and moving the volumes to the new VDISK.

How ever i don't know how to proceed. Can any one help please ?

P2000 Password Reset

$
0
0

Hi guys,

I have a P2000 storage and I have forgotten the password of P2000 storage. I couldn't log in both cli and ui. I had changed the password before a long time. I don't want to lose my datas, how to get recover my data without erasing?

I tried this information to login at the below,

admin\!admin

manage\!manage

but I couldn't still log in again

Have a nice day and thanks for your helping.

Which is the correct way to connect MSA2040 sas to HBA H421

$
0
0

Hi

I bought a MSA 2040 sas with 2 sas controllers (C8S53A), DL380 G9 whit 2 hba h241, autoloader and 3 sas cables, miniSas hd to minisas for the autoloader (Part NO: 716191-B21), and the others two cables are minisas hd to minisas hd for the msa to server (Part NO: 716195-B21) 

In theory always is compatible and should be works

The autoloader works fine whit her cable, but the msa connected to the server doesn't work with minisas hd cables, only works if I connect it using minisas hd to miniSas cable, I have tried some combinations with the 2 minisas HD cables but no one works

Please help!

 

MSA P2000 vdisk crash after replacing failed drive

$
0
0

Two days ago I began receiving emails from my P2000 that the drive in slot 12 was experiencing a unrecoverable read error (event 58).  This continued for a few minutes until the drive failed completely and the vdisk went into a critical state.

The next morning, I replaced the drive with a new one and configured it as a hot spare.  The vdisk began reconstruction, but after about 10 minutes, I received a vdisk quarantined message followed by a reconstruct failed message.

Following that the LED on the replacement drive in slot 12 stayed green, but the drive in slot 10 went amber and was labelled "leftovr" in the management utility. The vdisk was again in a critical state and was inaccessible, crashing several of my VMs.

I shut down all of my VMs and then the SAN, removed the new drive from slot 12 and restarted everything.  When everything was back up, the vdisk was again accessible though in a critical state being one drive short.

I repeated the drive replacement this morning. Cleared the metadata from the drive and added it as a hot spare. Again, in about 10 minute of reconstruction, the vdisk failed and the drive in slot 10 was marked "leftovr".  I removed the drive from slot 12 and the vdisk recovered in about 30 seconds.

Can anyone tell me what's going on?

MSA/SAN/Raid Performance

$
0
0
I have been looking for quite some time for information on trying to accuratly guage system/IO performance as it realtes to disk I/O. Trying to estimate performance differences between raid levels, drive speed etc. In addition to this judging what system I/O should be and measure it against what we are seeing with testing tools. Currently I believe we are getting HORRIBLE performance from our MSA1000 and I do not believe it is the product or the like but possibly how we have it configured. Does anyone have any REALLY GOOD information or links that could help me help myself?
Viewing all 2299 articles
Browse latest View live