HP 60 HP StorageWorks 60 Modular Smart Array Enclosure User Guide - Page 32

Failure of another drive during rebuild, Drive failure in a NetWare environment

Page 32 highlights

• The amount of unused capacity on the drives • The number of drives in the array (for RAID 5 and RAID 6 with ADG) Allow approximately 15 minutes per gigabyte for the rebuild process to be completed. This figure is conservative, and newer drive models usually require less time to rebuild. System performance is affected during the rebuild, and the system is unprotected against further drive failure until the rebuild has finished. Therefore, replace drives during periods of low activity when possible. CAUTION: If the Online LED of the replacement drive stops blinking and the amber Fault LED glows, or if other drive LEDs in the array go out, the replacement drive has failed and is producing unrecoverable disk errors. Remove and replace the failed replacement drive. When automatic data recovery has finished, the Online LED of the replacement drive stops blinking and begins to glow steadily. Failure of another drive during rebuild If a non-correctable read error occurs on another physical drive in the array during the rebuild process, the Online LED of the replacement drive stops blinking and the rebuild abnormally terminates. If this situation occurs, reboot the server. The system may temporarily become operational long enough to allow recovery of unsaved data. In any case, locate the faulty drive, replace it, and restore data from backup. Drive failure in a NetWare environment Use CPQONLIN to identify and monitor drive failure status in a NetWare environment. Failed drives or interim recovery mode If a drive fails and hardware fault tolerance is enabled, operation continues. Do the following: 1. Replace the drive as soon as possible. 2. Select a logical drive. 3. Press the F3 key to monitor to the status of drive recovery. Drive status messages include: • Interim Recovery: The logical drive is operating, but a failed drive has not been replaced. Replace the drive as soon as possible. • Ready for Recovery: The logical drives are queued for recovery. This status is displayed when another logical drive is already rebuilding or expanding. • Rebuilding: The array is operating and rebuilding a replacement drive or an online spare, if one was assigned. • Logical Drive Failed: If you have one or more logical drives that are not protected by fault tolerance in an array, the data on these logical drives will be lost. ACU shows the logical drives as FAILED. After drive replacement, any fault-tolerant logical drives rebuild. The logical drives that were not protected (FAILED) become available for data (the devices are reactivated automatically). If you have a backup of the data, restore the data now. If you do not replace the failed drive, the only option, using ACU, is to delete logical drives. Do not delete logical drives that contain valid data. Doing so results in data loss. NOTE: A failed status can occur on drives protected by fault tolerance if two or more physical drives fail concurrently. Troubleshooting 32

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52

Troubleshooting 32
The amount of unused capacity on the drives
The number of drives in the array (for RAID 5 and RAID 6 with ADG)
Allow approximately 15 minutes per gigabyte for the rebuild process to be completed. This figure is
conservative, and newer drive models usually require less time to rebuild.
System performance is affected during the rebuild, and the system is unprotected against further drive
failure until the rebuild has finished. Therefore, replace drives during periods of low activity when
possible.
CAUTION:
If the Online LED of the replacement drive stops blinking and the amber Fault LED glows, or if
other drive LEDs in the array go out, the replacement drive has failed and is producing unrecoverable disk
errors. Remove and replace the failed replacement drive.
When automatic data recovery has finished, the Online LED of the replacement drive stops blinking and
begins to glow steadily.
Failure of another drive during rebuild
If a non-correctable read error occurs on another physical drive in the array during the rebuild process,
the Online LED of the replacement drive stops blinking and the rebuild abnormally terminates.
If this situation occurs, reboot the server. The system may temporarily become operational long enough to
allow recovery of unsaved data. In any case, locate the faulty drive, replace it, and restore data from
backup.
Drive failure in a NetWare environment
Use CPQONLIN to identify and monitor drive failure status in a NetWare environment.
Failed drives or interim recovery mode
If a drive fails and hardware fault tolerance is enabled, operation continues. Do the following:
1.
Replace the drive as soon as possible.
2.
Select a logical drive.
3.
Press the
F3
key to monitor to the status of drive recovery.
Drive status messages include:
Interim Recovery
: The logical drive is operating, but a failed drive has not been replaced.
Replace the drive as soon as possible.
Ready for Recovery
: The logical drives are queued for recovery. This status is displayed when
another logical drive is already rebuilding or expanding.
Rebuilding
: The array is operating and rebuilding a replacement drive or an online spare, if one
was assigned.
Logical Drive Failed
: If you have one or more logical drives that are not protected by fault
tolerance in an array, the data on these logical drives will be lost. ACU shows the logical drives as
FAILED. After drive replacement, any fault-tolerant logical drives rebuild. The logical drives that were
not protected (FAILED) become available for data (the devices are reactivated automatically). If you
have a backup of the data, restore the data now.
If you do not replace the failed drive, the only option, using ACU, is to delete logical drives. Do
not
delete logical drives that contain valid data. Doing so results in data loss.
NOTE:
A failed status can occur on drives protected by fault tolerance if two or more physical drives fail
concurrently.