ISO8859-1@=`B % ( G ` y7F5K5 <! "#5$EQ(;)=*C+7U, 2N   -6 9 @%a(7)* + , $-. 5.$ d0 1 2' 3) 4% 6 37 K - kf -;)% !G ] RmTbn&l9<^`> b* ,3"-$4?'C(tD')vE*F+Q,?U2,Vo.%W.X/Y0LZ1 [1\22]y3^3_34t`e4a5bq7cc8_d/9Unable to open the debug file. The debug file must be opened prior to any write operations. An error has occurred during a print operation to the debug file. Unable to close the debug file. Unable to open message catalog file. Unable to initialize the ODM. Unable to lock the ODM. Unable to open the ODM. Unable to terminate the ODM. Unable to unlock the ODM. Unable to close the ODM. An error has occured during the odm_get_obj operation. The GDAM has been invoked on a system that does not contain an array. Unable to create a subprocess for system management. Unable to assign a closure function to a subprocess. Unable to assign an Exit Callback function to a subprocess. Unable to execute a subprocess. Unable to delete a subprocess. Unable to allocate memory. The ODM menus contains more option values than are allowed (MAX=20). Drives must be selected before using the drive pop-up menu.Devices must be selected before using the device pop-up menu.Subsystems must be selected before using the subsystem pop-up menu.A non-numeric character was entered in a numeric field.A required field was left empty.No command line options are supported by the IBM Graphical Disk Array Manager.OptionsRefreshExitHelpAbout...On Window...ContinueOkCancelIBM Graphical Disk Array ManagerIBM Graphical Disk Array Manager HelpRefresh in ProgressSet Refresh FrequencyWaitReading configuration. Display refresh in progress....Refresh frequency (seconds)Not foundCommand string:Command results:IBM Graphical Disk Array Manager - Version 1.1Licensed Materials - Property of IBM(C) Copyright IBM Corp. 1995All Rights Reserved.US Government Users Restricted Rights -Use, duplication or disclosure restrictedby GSA ADP Schedule Contract with IBMDefault value not foundFilling default field values... IBM Graphical Disk Array Manager V1.1 (C) Copyright IBM Corp. 1995 The IBM Graphical Disk Array Manager is a tool for creating, servicing and maintaining installed RAID disk array subsystems. The three subwindows show three categories of devices: 1) Subsystems, controllers and host adapters. Bus connections are drawn to show how subsystems are connected to the host. 2) Drives. These are the physical disk devices within the array subsystems. 3) Devices. These are logical disk devices (hdisks) within the array subsystems. To perform an action, first select one or more objects in a subwindow with the left mouse button. Then press the right mouse button within the subwindow to pop up a menu of possible actions. After selecting an action, a dialog window will be displayed that will let the user set parameters and confirm the operation. Note: The size of the subwindows can be changed by dragging the sash buttons on the right edge of the subwindow border. The entire window can also be resized by dragging the window borders. The subsystem window contains buttons for each adapter and subsystem connected to the host. Subsystems can be selected by positioning the mouse pointer over the host and pressing the left mouse button. The popup menu for subsystems can be invoked by pressing the right mouse button inside of the subsystem window. The drive window contains buttons for each physical drive in the selected subsystem. Drives can be selected by positioning the mouse pointer over the drive and pressing the left mouse button. The popup menu for drives can be invoked by pressing the right mouse button inside of the drive window. The device window contains buttons for each LUN in the selected subsystem. LUNs can be selected by positioning the mouse pointer over the host and pressing the left mouse button. The popup menu for subsystems can be invoked by pressing the right mouse button inside of the subsystem window. The refresh frequency is the amount of time between updates of the Graphical Disk Array Manager windows. The granularity of the time is in seconds. The time may be in the range of 1 to 9,999 seconds. The timer starts when the previous refresh completes. The Set Refresh Frequency function allows the user to configure the frequency that the display is updated to match the hardware configuration. When the time entered in the frequency field expires, the hardware configuration and status are read again to keep the display consistent with the state of the hardware. This makes the Graphical Disk Array Manager useful for monitoring array status. This is the device name of the selected 7135 RAIDiant Array subsystem. This is the amount of data the array controller reconstructs at a time. The recommended value is 256 blocks. The maximum value is 9,999. You can change the blocks to reconstruct per delay interval while the array controller is reconstructing data on a drive. This is useful if you want to vary the performance of the array during reconstuction. This field is used to customize the depth of the command queue for this logical unit. The range of values for this field are 4 to 30. The default value is 30.Selection of this option allows you to display/modify parameters associated with a specific RAIDiant Array. This is the specific logical unit which was selected on the prior selection panel. This is the block size (in bytes) which will be used by the controller to send data to the host system. The drive block size is 512 bytes. The block size for RAID levels 0,1 and 5 is 512. The block size for RAID level 3 is 1024 bytes for 3 drives and 2048 bytes for 5 drives. Note that in later versions, a block size of 512 will be used for all RAID levels. This is the size of the logical unit in megabytes. The combination of the RAID level and number of drives you specify determines the total number of megabytes in the logical unit. For example, a 5-drive RAID 5 logical unit with 2000 MByte drives (unformatted drive size) has a logical unit size of 8000 Mbytes. (Four drives are used for data and one for parity in a RAID 5 logical unit (4 X 2000 MByte = 8000 Mbyte). This is the number of LUN blocks contained in a segment (other than segment 0). This is the number of contiguous blocks which may be written to a single drive before switching to the next drive in the logical unit. It provides the host with a method to tune data striping to the typical IO request size. The controller determines a minimum and maximum segment size based on available memory and RAID level and adjusts the requested value accordingly. The default segment size is 512 blocks for RAID 0,1 and 5, and 1 block for RAID 3. The maximum segment size that the array controller accepts is 65,535 blocks. The first segment on a drive is segment zero. The default segment zero size is zero (0) blocks. If you change the value of segment zero to anything but zero, the array segments may not be properly aligned on the drives and array performance may decrease. This is the RAID level of the logical unit selected above. This can not be changed. The RAID level of a sub-LUN must be the same as the existing logical unit. This is the block size of the selected logical unit. This can not be changed from this menu. This is the segment size of the selected logical unit. This can not be changed from this menu. This is the segment 0 size of the selected logical unit. This can not be changed from this menu. This is the amount of time between data reconstruction operations. During this time, the array controller performs I/O system operation. As the delay interval increases, the system performance increases. The recommended value is one (1) tenth of a second. The maximum value allowed is 255 (or approximately 25 seconds). You can change the delay interval while the array controller is reconstructing data on a drive. This is useful if you want to vary the performance of the array during reconstruction. Use the Delete SCSI RAID Array option to delete logical units when you want to do the following: - Change the RAID level of an existing logical unit - Change the drives in an existing logical unit - Change the block size, array size, segment size or segment 0 size of an existing logical unit Before deleting the logical unit, you must backup all data in all logical volumes and filesystems on the logical unit and remove the logical unit from its volume group. Once the logical unit is deleted, you can NOT access the logical unit - the data is lost.Use this option to check and repair array parity. Array parity is used in RAID Levels 1, 3 and 5 logical units to enable data to be reconstructed if a single drive fails. Checking and repairing parity helps insure that you will be able to recover after a drive failure. You should check and repair parity after an abnormal system shutdown. Guaranteed Parity Scanning will be initially enabled. This means that a parity check will automatically occur at the following times: 1. At Start of Day 2. After a board reset if the reset occurred for any of the following reasons: a. change in redundant controller operational modes with forced quiescence. b. change in redundant controller operational roles (Passive to Active) with forced quiescence. c. microcode upload or download with forced quiescence. d. reset initiated over the board serial port Selection of this option will allow you to switch a particular disk group's LUN ownership from the selected Disk Array Controller to the other Disk Array Controller on the subsystem. This allows the user to set the frequency at which health checks will occur. Health checks will be performed on passive controllers and on active controllers which have no current LUN assignments. Allowable values for this field are 1 to 9,999 seconds. This allows the user to set the frequency at which polled Automatic Error Notification (AEN) request sense commands are issued to the controllers in the selected subsytem. Any detected errors will be logged by the system error logger. Allowable values for this field are 1 to 9,999 seconds. Selection of this option will provide dynamic load balancing in a dual active environment by monitoring the I/O request or data transfer rates and switching LUN ownership between Disk Array Controllers to achieve optimum performance. If the load balancing feature is enabled, this allows the user to set the frequency at which load balancing will occur. Allowable values for this field are 1 to 9,999 seconds. If 'yes' is selected, the logical unit will hold a lock on reservations attempted by other hosts. If 'no' is selected, other hosts will be able to utilize this logical unit. Typically, the user will mark a drive operational when adding a new drive to the array. This will change the 'nonexistent' drive to a 'spare' drive. You may also change a drive's status from 'warning' or 'failed' to 'operational', however, if done incorrectly, this may result in permanent loss of data. The option to mark a drive 'spare' should only be used to return a drive's status from 'hot spare' to 'spare'. The option to mark a drive 'hot spare' should be used only when the drive is currently marked as 'spare. This should only be done in a RAID 5 environment. For more information on the use of the 'hot spare' option, please refer to the IBM 7135 publications. The option to mark a drive 'replaced' is used to notify the RAIDiant array controller that you have physically replaced a 'failed' drive and that data reconstruction should begin. The option to 'format' a drive should only be used to force a format of an individual drive. This is generally not recommended and should instead be initiated via the IBM 7135 Service Aids. The option to 'fail' a drive is used to notify the controller that the particular drive is bad. The controller will then respond by taking the same steps it would have taken had it detected the failure. The option to 'delete' a drive should only be used to remove knowledge of its existence from the RAIDiant array controller. The drive will receive a 'non-existent' status as a result. Note that a drive which is part of an existing array may not be 'deleted'. These are the drives selected for this operation. This selected disk group's ownership will be switched to the other Disk Array Controller associated with this subsystem. This field lists the spare drives which will become associated with the new SCSI RAID array. A spare drive is a drive which exists within the RAIDiant array subsystem but which is not configured into an existing logical unit (LUN). This is the size of the logical unit in megabytes. This is the RAIDiant array controller which currently has LUN ownership of the selected drive group. The guidelines for RAID level selection are as follows: RAID 0 - Number of drives allowed = 1-10 RAID 1 - Number of drives allowed = 2,4,6,8 or 10 (The mirrored pair is created by grouping the first and second drive you enter, the third and fourth, etc. Drives in a mirrored pair cannot be on the same channel) RAID 3 - Number of drives allowed = 3 or 5 Drives must start on SCSI Channel 1. Drives must NOT skip channel numbers. The last drive entered is the parity drive. Note that some product versions will not allow RAID level 3. Refer to your specific product publication for this information. RAID 5 - Number of drives allowed = 3 - 5 This is the device name of the controller that will own the drive group after the logical unit has been created.This function creates a new logical unit (LUN) on a group of spare physical drives. The logical unit will be visible to the host system as a unique hdisk device. Typical OS commands can then be used to add the new logical unit to volume groups, logical volumes and filesystems (except in the case of RAID 3, which can only be accessed in raw mode).This function creates a new logical unit (LUN) using the same set of drives and RAID level as the existing RAIDiant array. You should create more than one logical unit on a single drive group only if creating a single logical unit would exceed the maximum logical unit size your system can support.