Antares-RAID-sparcLinux-HOWTO
 Thom Coates ([email protected]), Carl Munio, Jim Ludemann
 v0.1, 28 April 2000

 This document describes how to install, configure, and maintain a
 hardware RAID built around the 5070 SBUS host based RAID controller by
 Antares Microsystems.  Other topics of discussion include RAID levels,
 the 5070 controller GUI, and 5070 command line. A complete command
 reference for the 5070's K9 kernel and Bourne-like shell is included.
 ______________________________________________________________________

 Table of Contents



 1. Preamble

 2. Acknowledgements and Thanks

 3. New Versions

 4. Introduction

    4.1 5070 Main Features

 5. Background

    5.1 Raid Levels
    5.2 RAID Linear
             5.2...1 SUMMARY
    5.3 Level 1
             5.3...1 SUMMARY
    5.4 Striping
    5.5 Level 0
             5.5...1 SUMMARY:
    5.6 Level 2 and 3
             5.6...1 SUMMARY
    5.7 Level 4
             5.7...1 SUMMARY
    5.8 Level 5
             5.8...1 SUMMARY

 6. Installation

    6.1 SBUS Controller Compatibility
    6.2 Hardware Installation Procedure
             6.2...1 GNOME:
             6.2...2 KDE:
             6.2...3 XDM:
             6.2...4 Console Login (systems without X windows):
             6.2...5 All Systems:
             6.2...6 SPARCstation 4, 5, 10, 20 & UltraSPARC Systems:
             6.2...7 Ultra Enterprise Servers, SPARCserver 1000 & 2000 Systems, SPARCserver  6XO MP Series:
             6.2...8 All Systems:
             6.2...9 Verifying the Hardware Installation:
    6.3 Serial Terminal
    6.4 Hard Drive Plant

 7. 5070 Onboard Configuration

    7.1 Main Screen Options
             7.1...1 <Figure 1: Main Screen>
    7.2 [Q]uit
    7.3 [R]aidSets:
             7.3...1 <Figure 2: RAIDSet Configuration Screen>
    7.4 [H]ostports:
             7.4...1 <Figure 3: Host Port Configuration Screen>
    7.5 [S]pares:
             7.5...1 <Figure 4: Spare Device Configuration Screen>
    7.6 [M]onitor:
             7.6...1 <Figure 5: SCSI Monitor Screen>
    7.7 [G]eneral:
             7.7...1 <Figure 6: General Screen>
    7.8 [P]robe
    7.9 Example RAID Configuration Session

 8. Linux Configuration

    8.1 Existing Linux Installation
       8.1.1 QLogic SCSI Driver
       8.1.2 Device mappings
       8.1.3 Partitioning
       8.1.4 Installing a filesystem
       8.1.5 Mounting
    8.2 New Linux Installation

 9. Maintenance

    9.1 Activating a spare
    9.2 Re-integrating a repaired drive into the RAID (levels 3 and 5)

 10. Troubleshooting / Error Messages

    10.1 Out of band temperature detected...
    10.2 ... failed ... cannot have more than 1 faulty backend.
    10.3 When booting I see: ... Sun disklabel: bad magic 0000 ... unknown partition  table.

 11. Bugs

 12. Frequently Asked Questions

    12.1 How do I reset/erase the onboard configuration?
    12.2 How can I tell if a drive in my RAID has failed?

 13. Advanced Topics: 5070 Command Reference

    13.1 AUTOBOOT - script to automatically create all raid sets and scsi monitors
    13.2 AUTOFAULT - script to automatically mark a backend faulty after a drive  failure
    13.3 AUTOREPAIR - script to automatically allocate a spare and reconstruct a  raid set
    13.4 BIND - combine elements of the namespace
    13.5 BUZZER - get the state or turn on or off the buzzer
    13.6 CACHE - display information about and delete cache ranges
    13.7 CACHEDUMP - Dump the contents of the write cache to battery backed-up ram
    13.8 CACHERESTORE - Load the cache with data from battery backed-up ram
    13.9 CAT - concatenate files and print on the standard output
    13.10 CMP - compare the contents of 2 files
    13.11 CONS - console device for Husky
    13.12 DD - copy a file (disk, etc)
    13.13 DEVSCMP - Compare a file's size against a given value
    13.14 DFORMAT- Perform formatting functions on a backend disk drive
    13.15 DIAGS - script to run a diagnostic on a given device
    13.16 DPART - edit a scsihd disk partition table
    13.17 DUP - open file descriptor device
    13.18 ECHO - display a line of text
    13.19 ENV- environment variables file system
    13.20 ENVIRON - RaidRunner Global environment variables - names and effects
    13.21 EXEC - cause arguments to be executed in place of this shell
    13.22 EXIT - exit a K9 process
    13.23 EXPR - evaluation of numeric expressions
    13.24 FALSE - returns the K9 false status
    13.25 FIFO - bi-directional fifo buffer of fixed size
    13.26 GET - select one value from list
    13.27 GETIV - get the value an internal RaidRunner variable
    13.28 HELP - print a list of commands and their synopses
    13.29 HUSKY - shell for K9 kernel
    13.30 HWCONF - print various hardware configuration details
    13.31 HWMON - monitoring daemon for temperature, fans, PSUs.
    13.32 INTERNALS - Internal variables used by RaidRunner to change dynamics of  running kernel
    13.33 KILL - send a signal to the nominated process
    13.34 LED- turn on/off LED's on RaidRunner
    13.35 LFLASH- flash a led on RaidRunner
    13.36 LINE - copies one line of standard input to standard output
    13.37 LLENGTH - return the number of elements in the given list
    13.38 LOG - like zero with additional logging of accesses
    13.39 LRANGE - extract a range of elements from the given list
    13.40 LS - list the files in a directory
    13.41 LSEARCH - find the a pattern in a list
    13.42 LSUBSTR - replace a character in all elements of a list
    13.43 MEM - memory mapped file (system)
    13.44 MDEBUG - exercise and display statistics about memory allocation
    13.45 MKDIR - create directory (or directories)
    13.46 MKDISKFS - script to create a disk filesystem
    13.47 MKHOSTFS - script to create a host port filesystem
    13.48 MKRAID - script to create a raid given a line of output of rconf
    13.49 MKRAIDFS - script to create a raid filesystem
    13.50 MKSMON - script to start the scsi monitor daemon smon


 ______________________________________________________________________

 1.  Preamble


 Copyright 2000 by Thomas D. Coates, Jr. This document's source is
 licensed under the terms if the GNU general public license agreement.
 Permission to use, copy, modify, and distribute this document without
 fee for any purpose commercial or non-commercial is hereby granted,
 provided that the author's names and this notice appear in all copies
 and/or supporting documents; and that the location where a freely
 available unmodified version of this document may be obtained is
 given. This document is distributed in the hope that it will be
 useful, but WITHOUT ANY WARRANTY, either expressed or implied. While
 every effort has been taken to ensure the accuracy of the information
 documented herein, the
 author(s)/editor(s)/maintainer(s)/contributor(s) assumes NO
 RESPONSIBILITY for any errors, or for any damages, direct or
 consequential, as a result of the use of the information documented
 herein. A complete copy of the GNU Public License agreement may be
 obtained from: Free Software Foundation, Inc., 59 Temple Place - Suite
 330, Boston, MA 02111-1307, USA.  Portions of this document are
 adapted and/or re-printed from the 5070 installation guide and man
 pages with permission of Antares Microsystems, Inc., Campbell CA.

 2.  Acknowledgements and Thanks


 �  Carl and Jim at Antares for the hardware, man pages, and other
    support/contributions they provided during the writing of this
    document.

 �  Penn State University - Hershey Medical Center, Department of
    Radiology, Section of Clinical Image Management (My home away from
    my home away from home).

 �  The software-raid-HOWTO Copyright 1997 by Linas Vepstas under the
    GNU public license agreement. The software-raid-HOWTO is Available
    from : http://www.linuxdoc.org

 3.  New Versions


 �  The most recent version of this document can be found at my
    homepage: http://www.xray.hmc.psu.edu/~tcoates/

 �  Other versions may be found in different formats at the LDP
    homepage: http://www.linuxdoc.org and mirror sites.

 4.  Introduction


 The Antares 5070 is a high performance, versatile, yet relatively
 inexpensive host based RAID controller. Its embedded operating system
 (K9 kernel) is modelled on the Plan 9 operating system whose design is
 discussed in several papers from AT&T (see the "Further Reading"
 section). K9 is a kernel targeted at embedded controllers of small to
 medium complexity (e.g. ISDN-ethernet bridges, RAID controllers, etc).
 It supports multiple lightweight processes (i.e. without memory
 management) on a single CPU with a non-pre-emptive scheduler.  Device
 driver architecture is based on Plan 9 (and Unix SVR4) streams.
 Concurrency control mechanisms include semaphores and signals.


 The 5070 has three single ended ultra 1 SCSI channels and two onboard
 serial interfaces one of which provides command line access via a
 connected serial terminal or modem. The other is used to upgrade the
 firmware. The command line is robust, implementing many of the
 essential Unix commands (e.g. dd, ls, cat, etc.) and a scaled down
 Bourne shell for scripting. The Unix command set is augmented with
 RAID specific configuration commands and scripts. In addition to the
 command line interface an ASCII text based GUI is provided to permit
 easy configuration of level 0, 1, 3, 4, and 5 RAIDs.

 4.1.  5070 Main Features


 �  RAID levels 0, 1, 3, 4, and 5 are supported.

 �  Text based GUI for easy configuration for all supported RAID
    levels.

 �  A Multidisk RAID volume appears as an individual SCSI drive to the
    operating system and can be managed with the standard utilities
    (fdisk, mkfs, fsck,etc.).  RAID Volumes may be assigned to
    different SCSI IDs or the same SCSI IDs but different LUNs.

 �  No special RAID drivers required for the host operating system.

 �  Multiple RAID volumes of different levels can be mixed among the
    drives forming the physical plant. For example in a hypothetical
    drive plant consisting of 9 drives:

 �  2 drives form a level 3 RAID assigned to SCSI ID 5, LUN 0

 �  2 drives form a level 0 RAID assigned to SCSI ID 5, LUN 1

 �  5 drives form a level 5 RAID assigned to SCSI ID 6, LUN 0

 �  Three single ended SCSI channels which can accommodate 6 drives
    each (18 drives total).

 �  Two serial interfaces. The first permits
    configuration/control/monitoring of the RAID from a local serial
    terminal. The second serial port is used to upload new programming
    into the 5070 (using PPP and TFTP).

 �  Robust Unix-like command line and NVRAM based file system.

 �  Configurable ASCII SCSI communication channel for passing commands
    to the 5070's command line interpreter. Allows programming running
    on host OS to directly configure/control/monitor all parameters of
    the 5070.

 5.  Background


 Much of the information/knowledge pertaining to RAID levels in this
 section is adapted from the software-raid-HOWTO by Linas Vepstas . See
 the acknowledgements section for the URL where the full document may
 be obtained.
 RAID is an acronym for "Redundant Array of Inexpensive Disks" and is
 used to create large, reliable disk storage systems out of individual
 hard disk drives. There are two basic ways of implementing a RAID,
 software or hardware. The main advantage of a software RAID is low
 cost. However, since the OS of the host system must manage the RAID
 directly there is a substantial penalty in performance. Furthermore if
 the RAID is also the boot device, a drive failure could prove
 disastrous since the operating system and utility software needed to
 perform the recovery is located on the RAID. The primary advantages of
 hardware RAID is performance and improved reliability. Since all RAID
 operations are handled by a dedicated CPU on the controller, the host
 system's CPU is never bothered with RAID related tasks. In fact the
 host OS is completely oblivious to the fact that its SCSI drives are
 really virtual RAID drives. When a drive fails on the 5070 it can be
 replaced on-the-fly with a drive from the spares pool and its data
 reconstructed without the host's OS ever knowing anything has
 happened.

 5.1.  Raid Levels


 The different RAID levels have different performance, redundancy,
 storage capacity, reliability and cost characteristics. Most, but not
 all levels of RAID offer redundancy against drive failure. There are
 many different levels of RAID which have been defined by various
 vendors and researchers. The following describes the first 7 RAID
 levels in the context of the Antares 5070 hardware RAID
 implementation.

 5.2.  RAID Linear


 RAID-linear is a simple concatenation of drives to create a larger
 virtual drive. It is handy if you have a number small drives, and wish
 to create a single, large drive. This concatenation offers no
 redundancy, and in fact decreases the overall reliability: if any one
 drive fails, the combined drive will fail.

 5.2.0.0.1.  SUMMARY


 �  Enables construction of a large virtual drive from a number of
    smaller drives

 �  No protection, less reliable than a single drive

 �  RAID 0 is a better choice due to better I/O performance

 5.3.  Level 1


 Also referred to as "mirroring". Two (or more) drives, all of the same
 size, each store an exact copy of all data, disk-block by disk-block.
 Mirroring gives strong protection against drive failure: if one drive
 fails, there is another with the an exact copy of the same data.
 Mirroring can also help improve performance in I/O-laden systems, as
 read requests can be divided up between several drives. Unfortunately,
 mirroring is also one of the least efficient in terms of storage: two
 mirrored drives can store no more data than a single drive.

 5.3.0.0.1.  SUMMARY


 �  Good read/write performance


 �  Inefficient use of storage space (half the total space available
    for data)

 �  RAID 6 may be a better choice due to better I/O performance.

 5.4.  Striping


 Striping is the underlying concept behind all of the other RAID
 levels.  A stripe is a contiguous sequence of disk blocks. A stripe
 may be as short as a single disk block, or may consist of thousands.
 The RAID drivers split up their component drives into stripes; the
 different RAID levels differ in how they organize the stripes, and
 what data they put in them. The interplay between the size of the
 stripes, the typical size of files in the file system, and their
 location on the drive is what determines the overall performance of
 the RAID subsystem.

 5.5.  Level 0


 Similar to RAID-linear, except that the component drives are divided
 into stripes and then interleaved. Like RAID-linear, the result is a
 single larger virtual drive. Also like RAID-linear, it offers no
 redundancy, and therefore decreases overall reliability: a single
 drive failure will knock out the whole thing. However, the 5070
 hardware RAID 0 is the fastest of any of the schemes listed here.

 5.5.0.0.1.  SUMMARY:


 �  Use RAID 0 to combine smaller drives into one large virtual drive.

 �  Best Read/Write performance of all the schemes listed here.

 �  No protection from drive failure.

 �  ADVICE: Buy very reliable hard disk drives if you plan to use this
    scheme.

 5.6.  Level 2 and 3


 RAID-2 is seldom used anymore, and to some degree has been made
 obsolete by modern hard disk technology. RAID-2 is similar to RAID-4,
 but stores ECC information instead of parity. Since all modern disk
 drives incorporate ECC under the covers, this offers little additional
 protection. RAID-2 can offer greater data consistency if power is lost
 during a write; however, battery backup and a clean shutdown can offer
 the same benefits. RAID-3 is similar to RAID-4, except that it uses
 the smallest possible stripe size.

 5.6.0.0.1.  SUMMARY


 �  RAID 2 is largely obsolete

 �  Use RAID 3 to combine separate drives together into one large
    virtual drive.

 �  Protection against single drive failure,

 �  Good read/write performance.



 5.7.  Level 4


 RAID-4 interleaves stripes like RAID-0, but it requires an additional
 drive to store parity information. The parity is used to offer
 redundancy: if any one of the drives fail, the data on the remaining
 drives can be used to reconstruct the data that was on the failed
 drive. Given N data disks, and one parity disk, the parity stripe is
 computed by taking one stripe from each of the data disks, and XOR'ing
 them together. Thus, the storage capacity of a an (N+1)-disk RAID-4
 array is N, which is a lot better than mirroring (N+1) drives, and is
 almost as good as a RAID-0 setup for large N. Note that for N=1, where
 there is one data disk, and one parity disk, RAID-4 is a lot like
 mirroring, in that each of the two disks is a copy of each other.
 However, RAID-4 does NOT offer the read-performance of mirroring, and
 offers considerably degraded write performance.  In brief, this is
 because updating the parity requires a read of the old parity, before
 the new parity can be calculated and written out. In an environment
 with lots of writes, the parity disk can become a bottleneck, as each
 write must access the parity disk.

 5.7.0.0.1.  SUMMARY


 �  Similar to RAID 0

 �  Protection against single drive failure.

 �  Poorer I/O performance than RAID 3

 �  Less of the combined storage space is available for data [than RAID
    3] since an additional drive is needed for parity information.

 5.8.  Level 5


 RAID-5 avoids the write-bottleneck of RAID-4 by alternately storing
 the parity stripe on each of the drives. However, write performance is
 still not as good as for mirroring, as the parity stripe must still be
 read and XOR'ed before it is written. Read performance is also not as
 good as it is for mirroring, as, after all, there is only one copy of
 the data, not two or more. RAID-5's principle advantage over mirroring
 is that it offers redundancy and protection against single-drive
 failure, while offering far more storage capacity when used with three
 or more drives.

 5.8.0.0.1.  SUMMARY


 �  Use RAID 5 if you need to make the best use of your available
    storage space while gaining protection against single drive
    failure.

 �  Slower I/O performance than RAID 3

 6.  Installation


 NOTE: The installation procedure given here for the SBUS controller is
 similar to that found in the manual. It has been modified so minor
 variations in the SPARCLinux installation may be included.

 6.1.  SBUS Controller Compatibility



 The 5070 / Linux 2.2 combination was tested on SPARCstation (5, 10, &
 20), Ultra 1, and Ultra 2 Creator. The 5070 was also tested on Linux
 with Symmetrical Multiprocessing (SMP) support on a dual processor
 Ultra 2 creator 3D with no problems. Other 5070 / Linux / hardware
 combinations may work as well.

 6.2.  Hardware Installation Procedure


 If your system is already up and running, you must halt the operating
 system.

 6.2.0.0.1.  GNOME:


 1. From the login screen right click the "Options" button.

 2. On the popup menu select System -> Halt.

 3. Click "Yes" when the verification box appears


 6.2.0.0.2.  KDE:


 1. From the login screen right click shutdown.

 2. On the popup menu select shutdown by right clicking its radio
    button.

 3. Click OK


 6.2.0.0.3.  XDM:


 1. login as root

 2. Left click on the desktop to bring up the pop-up menu

 3. select "New Shell"

 4. When the shell opens type "halt" at the prompt and press return


 6.2.0.0.4.  Console Login (systems without X windows):


 1. Login as root

 2. Type "halt"


 6.2.0.0.5.  All Systems:


 Wait for the message "power down" or "system halted" before
 proceeding. Turn off your SPARCstation system (Note: Your system may
 have turned itself off following the power down directive), its video
 monitor, external disk expansion boxes, and any other peripherals
 connected to the system.  Be sure to check that the green power LED on
 the front of the system enclosure is not lit and that the fans inside
 the system are not running. Do not disconnect the system power cord.



 6.2.0.0.6.  SPARCstation 4, 5, 10, 20 & UltraSPARC Systems:


 1. Remove the top cover on the CPU enclosure. On a SPARCstation 10,
    this is done by loosening the captive screw at the top right corner
    of the back of the CPU enclosure, then tilting the top of the
    enclosure forward while using a Phillips screwdriver to press the
    plastic tab on the top left corner.

 2. Decide which SBUS slot you will use. Any slot will do. Remove the
    filler panel for that slot by removing the two screws and
    rectangular washers that hold it in.

 3. Remove the SBUS retainer (commonly called the handle) by pressing
    outward on one leg of the retainer while pulling it out of the hole
    in the printed circuit board.

 4. Insert the board into the SBUS slot you have chosen. To insert the
    board, first engage the top of the 5070 RAIDium backpanel into the
    backpanel of the CPU enclosure, then rotate the board into a level
    position and mate the SBUS connectors. Make sure that the SBUS
    connectors are completely engaged.

 5. Snap the nylon board retainers inside the SPARCstation over the
    5070 RAIDium board to secure it inside the system.

 6. Secure the 5070 RAIDium SBUS backpanel to the system by replacing
    the rectangular washers and screws that held the original filler
    panel in place.

 7. Replace the top cover by first mating the plastic hooks on the
    front of the cover to the chassis, then rotating the cover down
    over the unit until the plastic tab in back snaps into place.
    Tighten the captive screw on the upper right corner.


 6.2.0.0.7.  6XO MP Series: Ultra Enterprise Servers, SPARCserver 1000
 &amp; 2000 Systems, SPARCserver


 1. Remove the two Allen screws that secure the CPU board to the card
    cage.  These are located at each end of the CPU board backpanel.

 2. Remove the CPU board from the enclosure and place it on a static-
    free surface.

 3. Decide which SBUS slot you will use. Any slot will do. Remove the
    filler panel for that slot by removing the two screws and
    rectangular washers that hold it in. Save these screws and washers.

 4. Remove the SBUS retainer (commonly called the handle) by pressing
    outward on one leg of the retainer while pulling it out of the hole
    in the printed circuit board.

 5. Insert the board into the SBUS slot you have chosen. To insert the
    board, first engage the top of the 5070 RAIDium backpanel into the
    backpanel of the CPU enclosure, then rotate the board into a level
    position and mate the SBUS connectors. Make sure that the SBUS
    connectors are completely engaged.

 6. Secure the 5070 RAIDium board to the CPU board with the nylon
    screws and standoffs provided on the CPU board. The standoffs may
    have to be moved so that they match the holes used by the SBUS
    retainer, as the standoffs are used in different holes for an MBus
    module. Replace the screws and rectangular washers that originally
    held the filler panel in place, securing the 5070 RAIDium SBus
    backpanel to the system enclosure.

 7. Re-insert the CPU board into the CPU enclosure and re-install the
    Allen-head retaining screws that secure the CPU board.


 6.2.0.0.8.  All Systems:


 1. Mate the external cable adapter box to the 5070 RAIDium and gently
    tighten the two screws that extend through the cable adapter box.

 2. Connect the three cables from your SCSI devices to the three 68-pin
    SCSI-3 connectors on the Antares 5070 RAIDium. The three SCSI
    cables must always be reconnected in the same order after a RAID
    set has been established, so you should clearly mark the cables and
    disk enclosures for future disassembly and reassembly.

 3. Configure the attached SCSI devices to use SCSI target IDs other
    than 7, as that is taken by the 5070 RAIDium itself. Configuring
    the target number is done differently on various devices. Consult
    the manufacturer's installation instructions to determine the
    method appropriate for your device.

 4. As you are likely to be installing multiple SCSI devices, make sure
    that all SCSI buses are properly terminated. This means a
    terminator is installed only at each end of each SCSI bus daisy
    chain.


 6.2.0.0.9.  Verifying the Hardware Installation:


 These steps are optional but recommended. First, power-on your system
 and interrupt the booting process by pressing the "Stop" and "a" keys
 (or the "break" key if you are on a serial terminal) simultaneously as
 soon as the Solaris release number is shown on the screen. This will
 force the system to run the Forth Monitor in the system EPROM, which
 will display the "ok" prompt. This gives you access to many useful
 low-level commands, including:

 ok show-devs



 . . .



 /iommu@f,e0000000/sbus@f,e000100SUNW, isp@1,8800000



 . . .



 The first line in the response shown above means that the 5070 RAIDium
 host adapter has been properly recognized. If you don't see a line
 like this, you may have a hardware problem.


 Next, to see a listing of all the SCSI devices in your system, you can
 use the probe-scsi-all command, but first you must prepare your system
 as follows:

 ok setenv auto-boot? False



 ok reset



 ok probe-scsi-all



 This will tell you the type, target number, and logical unit number of
 every SCSI device recognized in your system. The 5070 RAIDium board
 will report itself attached to an ISP controller at target 0 with two
 Logical Unit Numbers (LUNs): 0 for the virtual hard disk drive, and 7
 for the connection to the Graphical User Interface (GUI). Note: the
 GUI communication channel on LUN 7 is currently unused under Linux.
 See the discussion under "SCSI Monitor Daemon (SMON)" in the "Advanced
 Topics" section for more information.


 REQUIRED: Perform a reconfiguration boot of the operating system:

 ok boot -r



 If no image appears on your screen within a minute, you most likely
 have a hardware installation problem. In this case, go back and check
 each step of the installation procedure. This completes the hardware
 installation procedure.


 6.3.  Serial Terminal


 If you have a serial terminal at your disposal (e.g. DEC-VT420) it may
 be connected to the controller's serial port using a 9 pin DIN male to
 DB25 male serial cable. Otherwise you will need to supplement the
 above cable with a null modem adapter to connect the RAID controller's
 serial port to the serial port on either the host computer or a PC.
 The terminal emulators I have successfully used include Minicom (on
 Linux), Kermit (on Caldera's Dr. DOS), and Hyperterminal (on a windows
 CE palmtop), however, any decent terminal emulation software should
 work. The basic settings are 9600 baud , no parity, 8 data bits, and 1
 stop bit.

 6.4.  Hard Drive Plant


 Choosing the brand and capacity of the drives that will form the hard
 drive physical plant is up to you. I do have some recommendations:

 �  Remember, you generally get what you pay for. I strongly recommend
    paying the extra money for better (i.e. more reliable) hardware
    especially if you are setting up a RAID for a mission critical
    project. For example, consider purchasing drive cabinets with
    redundant hot-swappable power supplies, etc.


 �  You will also want a UPS for your host system and drive cabinets.
    Remember, RAID levels 3 and 5 protect you from data loss due to
    drive failure NOT power failure.


 �  The drive cabinet you select should have hot swappable drive bays,
    these cost more but are definitely worth it when you need to
    add/change drives.

 �  Make sure the cabinet(s) have adequate cooling when fully loaded
    with drives.

 �  Keep your SCSI cables (internal and external) as short as possible

 �  Mark the drives/cabinet(s) in such a way that you will be able to
    reconnect them to the controller in their original configuration.
    Once the RAID is configured you cannot re-organize you drives
    without re-configuring the RAID (and subsequently erasing the data
    stored on it).

 �  Keep in mind that although it is physically possible to
    connect/configure up to 6 drives per channel, performance will
    sharply decrease for RAIDs with more than three drives per channel.
    This is due to the 25 MHz bandwidth limitation of the SBUS.
    Therefore, if read/write performance is an issue go with a small
    number of large drives. If you need a really large RAID (~ 1
    terabyte) then you will have no other choice but to load the
    channels to capacity and pay the performance penalty. NOTE: if you
    are serving files over a 10/100 Base T network you may not notice
    the performance decrease since the network is usually the
    bottleneck not the SBUS.

 7.  5070 Onboard Configuration


 Before diving into the RAID configuration I need to define a few
 terms.

 �  "RaidRunner" is the name given to the the 5070 controller board.

 �  "Husky" is the name given to the shell which produces the ":raid;"
    command prompt. It is a command language interpreter that executes
    commands read from the standard input or from a file. Husky is a
    scaled down model of Unix's Bourne shell (sh). One major difference
    is that husky has no concept of current working directory. For more
    information on the husky shell and command prompt see the "Advanced
    Topics" section

 �  The "host port" is the SCSI ID assigned to the controller card
    itself. This is usually ID 7.

 �  A "backend" is a drive attached to the controller on a given
    channel.

 �  A "rank" is a collection of all the backends from each channel with
    the same SCSI ID (i.e. rank 0 would consist of all the drives with
    SCSI ID 0 on each channel)

 �  Each of the backends is identified by a three digit number where
    the first digit is the channel, the second the SCSI ID of the
    drive, and the third the LUN of the drive. The numbers are
    separated by a period. The identifier is prefixed with a "D" if it
    is a disk or "T" if it is a tape (e.g. D0.1.0). This scheme is
    referred to as <device_type_c.s.l> in the following documentation.

 �  A "RAID set" consists of given number of backends (there are
    certain requirements which I'll come to later)

 �  A "spare" is a drive which is unused until there is a failure in
    one of the RAID drives. At that time the damaged drive is
    automatically taken offline and replaced with the spare. The data
    is then reconstructed on the spare and the RAID resumes normal
    operation.

 �  Spares may either be "hot" or "warm" depending on user
    configuration. Hot spares are spun up when the RAID is started,
    which shortens the replacement time when a drive failure occurs.
    Warm spares are spun up when needed, which saves wear on the drive.

 The test based GUI can be started by typing "agui"

 : raid; agui



 at the husky prompt on the serial terminal (or emulator).

 Agui is a simple ASCII based GUI that can be run on the RaidRunner
 console port which enables one to configure the RaidRunner. The only
 argument agui takes is the terminal type that is connected to the
 RaidRunner console. Current supported terminals are dtterm, vt100 and
 xterm. The default is dtterm.

 Each agui screen is split into two areas, data and menu. The data
 area, which generally uses all but the last line of the screen,
 displays the details of the information under consideration. The menu
 area, which generally is the bottom line of the screen, displays a
 strip menu with a title then list of options or sub-menus. Each option
 has one character enclosed in square brackets (e.g. [Q]uit) which is
 the character to type to select that option.  Each menu line allows
 you to refresh the screen data (in case another process on the
 RaidRunner writes to the console). The refresh character may also be
 used during data entry if the screen is overwritten. The refresh
 character is either <Control-l> or <Control-r>.

 When agui starts, it reads the configuration of the RaidRunner and
 probes for every possible backend. As it probes for each backend, it's
 "name" is displayed in the bottom left corner of the screen.

 7.1.  Main Screen Options


 7.1.0.0.1.  <Figure 1: Main Screen>


 The Main screen is the first screen displayed. It provides a summary
 of the RaidRunner configuration. At the top is the RaidRunner model,
 version and serial number. Next is a line displaying, for each
 controller, the SCSI ID's for each host port (labeled A, B, C, etc)
 and total and currently available amounts of memory. The next set of
 lines display the ranks of devices on the RaidRunner. Each device
 follows the nomenclature of <device_type_c.s.l> where device_type_ can
 be D for disk or T for tape, c is the internal channel the device is
 attached to, s is the SCSI ID (Rank) of the device on that channel,
 and l is the SCSI LUN of the device (typically 0).

 The next set of lines provide a summary of the Raid Sets configured on
 the RaidRunner. The summary includes the raid set name, it's type,
 it's size, the amount of cache allocated to it and a comma separated
 list of it's backends.  See rconf in the "Advanced Topics" section for
 a full description of the above.

 Next are the spare devices configured. Each spare is named
 (device_type_c.s.l format), followed by it's size (in 512-byte
 blocks), it's spin state (Hot or Warm), it's controller allocation ,
 and finally it's current status (Used/Unused, Faulty/Working). If
 used, the raid set that uses it is nominated.
 At the bottom of the data area, the number of controllers, channels,
 ranks and devices are displayed.

 The menu line allows one to quit agui or select further actions or
 sub-menus.

 �  [Q]uit: Exit the main screen and return to the husky prompt.

 �  [R]aidSets: Enter the RaidSet configuration screen.

 �  [H]ostports Enter the Host Port configuration screen.

 �  [S]pares Enter the Spare Device configuration screen.

 �  [M]onitor Enter the SCSI Monitor configuration screen.

 �  [G]eneral Enter the General configuration/information screen.

 �  [P]robe Re-probe the device backends on the RaidRunner. As each
    backend is probed it's "name" (c.s.l format) is displayed in the
    bottom left corner of the screen.

 These selections are described in detail below.

 7.2.  [Q]uit


 Exit the agui main screen and return to the husky ( :raid; ) prompt.

 7.3.  [R]aidSets:


 7.3.0.0.1.  <Figure 2: RAIDSet Configuration Screen>


 The Raid Set Configuration screen displays a Raid Set in the data area
 and provides a menu which allows you to Add, Delete, Modify, Install
 (changes) and Scroll through all other raid sets (First, Last, Next
 and Previous). If no raid sets have been configured, only the screen
 title and menu is displayed.  All attributes of the raid set are
 displayed. For information on each attribute of the raid set, see the
 rconf command in the "Advanced Topics" section. The menu line allows
 one to leave the Raid Set Configuration screen or select further
 actions:

 �  [Q]uit: Exit the Raid Set Configuration screen and return to the
    Main screen. If you have modified, deleted or added a raid set and
    have not installed the changes you will be asked to confirm this.
    If you select Yes to continue the exit, all changes made since the
    last install action will be discarded.

 �  [I]nst: This action installs (into the RaidRunner configuration
    area) any changes that may have been made to raid sets, be that
    deletion, addition or modification. If you exit prior to
    installing, all changes made since the last installation will be
    discarded. The installation process takes time. It is complete once
    the typed "i" character, is cleared from the menu line.

 �  [M]od: This action allows you to modify the displayed raid set.
    You will be prompted for each Raid Set attribute that can be
    changed. The prompt includes allowable options or formats required.
    If you don't wish to change a particular attribute, then press the
    RETURN or TAB key. The attributes you can change are the raid set
    name, I/O mode, status (Active to Inactive), bootmode, spares
    usage, backend zone table usage, IO size (if raid set has never
    been used - i.e. just added), cache size, I/O queues length, host
    interfaces and additional stargd arguments. If you wish to change a
    single attribute then use the RETURN or TAB key to skip all other
    options. The changed attribute will be re-displayed as soon as you
    press the RETURN key. When specifying cache size, you may suffix
    the number with 'm' or 'M' to indicate the number is in Megabytes
    or with 'k' or 'K' to indicate the number is in Kilobytes. Note you
    can only enter whole integer values. When specifying io size, you
    may suffix the number with 'k' or 'K' to indicate the number is in
    Kilobytes. When you enter data, it is checked for correctness and
    if incorrect, a message is displayed and all changes are discarded
    and you will have to start again. Remember you must install
    ([I]nst.) any changes.

 �  [A]dd: When this option is selected you will be prompted for
    various attributes of the new raid set. These attributes are the
    raid set name, the raid set type, the initial host interface the
    raid set is to appear on (in c.h.l format where c is the controller
    number, h is the host port (0, 1, 2 etc) and l is the SCSI LUN) and
    finally a list of backends. When backends are to be entered, the
    screen displays a list of available backends, each with a numeric
    index (commencing at 0). You select each backend by entering the
    index and once complete enter q for Quit. As each backend index is
    entered, it's backend name is displayed in a comma separated list.
    When you enter data, it is checked for correctness and if
    incorrect, a message is displayed and the addition will be ignored
    and you will have to start again. Once the backends are complete,
    the newly created raid set will be displayed on the screen with
    supplied and default attributes. You can then modify the raid set
    to change other attributes. Remember you must install ([I]nst.) any
    new raid sets.

 �  [D]elete: This action will delete the currently displayed raid set.
    If this raid set is Active, then you will not be allowed to delete
    it.  You will have to make it Inactive (via the [M]od. option) then
    delete it. You will be prompted to confirm the deletion. Once you
    confirm the deletion, the screen will be cleared and the next raid
    set will be displayed, if configured.  Remember you must install
    ([I]nst.) any changes.

 �  [F]irst, [L]ast, [N]ext and [P]rev allow you to scroll through the
    configured raid sets.

 7.4.  [H]ostports:


 7.4.0.0.1.  <Figure 3: Host Port Configuration Screen>


 The Host Port Configuration screen displays for each controller, each
 host port (labelled A, B, C, etc for port number 0, 1, 2, etc) and the
 assigned SCSI ID. If the RaidRunner you use, has external switches for
 host port SCSI ID selection, you may only exit ([Q]uit) from this
 screen. If the RaidRunner you use, does NOT have external switches for
 host port SCSI ID selection, then you may modify (and hence install)
 the SCSI ID for any host port. The menu line allows one to leave the
 Host Port Configuration screen or select further actions (if NO
 external host):

 �  [Q]uit: Exit the Host Port Configuration screen and return to the
    Main screen. If you have modified a host port SCSI ID assignment
    and have not installed the changes you will be asked to confirm
    this. If you select Yes to continue the exit, all changes made
    since the last install action will be discarded.

 �  [I]nstall: This action installs (into the RaidRunner configuration
    area) any changes that may have been made to host port SCSI ID
    assign� ments.  If you exit prior to installing, all changes made
    since the last installation will be discarded. The installation
    process takes time. It is complete once the typed "i" character, is
    cleared from the menu line.

 �  [M]odify: This action allows you to modify the host port SCSI ID
    assignments for each host port on each controller (if NO external
    host port SCSI ID switches). You will be prompted for the SCSI ID
    for each host port.  You can enter either a SCSI ID (0 thru 15),
    the minus "-" character to clear the SCSI ID assignment or RETURN
    to SKIP. As you enter data, it is checked for correctness and if
    incorrect, a message will be printed although previously correctly
    entered data will be retained. Remember you must install ([I]nst.)
    any changes.

 7.5.  [S]pares:


 7.5.0.0.1.  <Figure 4: Spare Device Configuration Screen>


 The Spare Device Configuration screen displays all configured spare
 devices in the data area and provides a menu which allows you to Add,
 Delete, Mod� ify and Install (changes) spare devices. If no spare
 devices have been configured, only the screen title and menu is
 displayed. Each spare device displayed, shows it's name (in
 device_type_c.s.l format), it's size in 512-byte blocks, it's spin
 status (Hot or Warm), it's controller allocation, finally it's current
 status (Used/Unused, Faulty/Working). If used, the raid set that uses
 it is nominated. For information on each attribute of a spare device,
 see the rconf command in the "Advanced Topics" section. The menu line
 allows one to leave the Spare Device Configuration screen or select
 further actions:

 �  [Q]uit: Exit the Spare Device Configuration screen and return to
    the Main screen. If you have modified, deleted or added a spare
    device and have not installed the changes you will be asked to
    confirm this. If you select Yes to continue the exit, all changes
    made since the last install action will be discarded.

 �  [I]nstall: This action installs (into the RaidRunner configuration
    area) any changes that may have been made to the spare devices, be
    that deletion, addition or modification. If you exit prior to
    installing, all changes made since the last installation will be
    discarded. The installation process takes time. It is complete once
    the typed "i" character, is cleared from the menu line.

 �  [M]odify: This action allows you to modify the unused spare
    devices.  You will be prompted for each spare device attribute that
    can be changed. The prompt includes allowable options or formats
    required. If you don't wish to change a particular attribute, then
    press the RETURN key. The attributes you can change are the new
    size (in 512-byte blocks), the spin state (H or hot or W for Warm),
    and the controller allocation (A for any, 0 for controller 0, 1 for
    controller 1, etc). If you wish to change a single attribute of a
    spare device, then use the RETURN key to skip all other attributes
    for each spare device. The changed attribute will not be re-
    displayed until the last prompted attribute is entered (or
    skipped). When you enter data, it is checked for cor� rectness and
    if incorrect, a message is dis� played and all changes are
    discarded and you will have to start again. Remember you must
    install ([I]nstall) any changes.

 �  [A]dd: When adding a spare device, the list of available devices is
    displayed and you are required to type in the device name. Once
    entered, the spare is added with defaults which you can change, if
    required, via the [M]odify option. Remember you must install
    ([I]nstall) any changes.

 �  [D]elete: When deleting a spare device, the list of spare devices
    allowed to be deleted is displayed and you are required to type in
    the required device name. Once entered, the spare is deleted from
    the screen. Remember you must install ([I]nstall) any changes.

 7.6.  [M]onitor:


 7.6.0.0.1.  <Figure 5: SCSI Monitor Screen>


 The SCSI Monitor Configuration screen displays a table of SCSI
 monitors configured for the RaidRunner. Up to four SCSI monitors may
 be configured.  The table columns are entitled Controller, Host Port,
 SCSI LUN and Protocol and each line of the table shows the appropriate
 SCSI Monitor attribute. For details on SCSI Monitor attributes, see
 the rconf command in the "Advanced Topics" section. The menu line
 allows one to leave the SCSI Monitor Configuration screen or modify
 and install the table.

 �  [Q]uit: Exit the SCSI Monitor Configuration screen and return to
    the Main screen. If you have made changes and have not installed
    them you will be asked to confirm this. If you select Yes to
    continue the exit, all changes made since the last install action
    will be discarded.

 �  [I]nstall: This action installs (into the RaidRunner configuration
    area) any changes that may have been made to SCSI Monitor
    configuration. If you exit prior to installing, all changes made
    since the last installation will be discarded. The installation
    process takes time. It is complete once the typed "i" character, is
    cleared from the menu line.

 �  [M]odify: This action allows you to modify the SCSI Monitor
    configuration.  The cursor will be moved around the table,
    prompting you for input. If you do not want to change an attribute,
    enter RETURN to skip. If you want to delete a SCSI monitor then
    enter the minus "-" character when prompted for the controller
    number. If you want to use the default protocol list, then enter
    RETURN at the Protocol List prompt. As you enter data, it is
    checked for correctness and if incorrect, a message will be printed
    and any previously entered data is discarded. You will have to re-
    enter the data again. Remember you must install ([I]nstall) any
    changes.

 7.7.  [G]eneral:


 7.7.0.0.1.  <Figure 6: General Screen>


 The General screen has a blank data area and a menu which allows one
 to Quit and return to the main screen, or to select further sub-menus
 which provide information about Devices, the System Message Logger,
 Global Environment variables and throughput Statistics.

 �  [Q]uit: Exit the General screen and return to the Main screen.

 �  [D]evices: Enter the Device information screen. The Devices screen
    displays the name of all devices on the RaidRunner. The menu line
    allows one to Quit and return to the General screen or display
    information about the devices.

    <Figure 7: Devices Screen>


 �  [Q]uit: Exit the Devices screen and return to the General screen.


 �  [I]nformation: The Device Information screen displays information
    about each device. You can scroll through the devices. For disks,
    information displayed includes, the device name, serial number,
    vendor name, product id, speed, version, sector size, sector count,
    total device size in MB, number of cylinders, heads and sectors per
    track and the zone/notch partitions. The menu line allows one the
    leave the Device Information screen or browse through devices.

    <Figure 8: Device Information Screen>


 �  [Q]uit: Exit the Device Information screen and return to the
    Devices screen.

 �  [F]irst, [L]ast, [N]ext and [P]rev allow you to scroll through the
    devices and hence display their current data .

 �  Sys[L]og: Enter the System Logger Messages screen.

    <Figure 9: System Logger Messages Screen>


 �  [Q]uit: Exit the System Logger Messages screen and return to the
    General screen.

 �  [F]irst, [L]ast, [N]ext and [P]rev allow you to scroll through the
    system log.

 �  [E]nvironment: Enter the Global Environment Variable configuration
    screen. The Environment Variable Configuration screen dis� plays
    all configured Global Environment Variables and provides a menu
    which allows you to Add, Delete, Modify and Install (changes)
    variables. Each variable name is displayed followed by an equals
    "=" and the value assigned to that variable enclosed in braces -
    "{" .. "}". The menu line allows you to Quit and return to the
    General screen or select further actions.

    <Figure 10: Environment Global Variable Configuration Screen>


 �  [Q]uit: Exit the Environment Variable Configuration screen and
    return to the General screen. If you have modified, deleted or
    added an environment variable and have not installed the changes
    you will be asked to confirm this.  If you select Yes to continue
    the exit, all changes made since the last install action will be
    discarded.

 �  [I]nst: This action installs (into the RaidRunner configuration
    area) any changes that may have been made to environment variables,
    be that deletion, addition or modification. If you exit prior to
    installing, all changes made since the last installation will be
    discarded. The installation process takes time. It is complete once
    the typed "i" character, is cleared from the menu line.

 �  [M]od: This action allows you to modify an environment variable's
    value. You will be prompted for the name of the environment
    variable and then prompted for it's new value. If the environment
    variable entered is not found, a message will be printed and you
    will not be prompted for a new value. If you do not enter a new
    value, (i.e. just press RETURN) no change will be made.  Remember
    you must install ([I]nstall) any changes.

 �  [A]dd: When adding a new environment variable, you will be prompted
    for it's name and value. Providing the variable name is not already
    used and you enter a value, the new variable will be added and
    displayed. Remember you must install ([I]nstall) any changes.

 �  [D]elete: When deleting an environment variable, you will be
    prompted for the variable name and if valid, the environment
    variable will be deleted. Remember you must install ([I]nstall) any
    changes.

 �  [S]tats: Enter the Statistics monitoring screen. The Statistics
    screen display various general and specific statistics about raid
    sets configured and running on the RaidRunner. The first section of
    the data area displays the current temperature in degrees Celsius
    and the current speed of fans in the RaidRunner. The next section
    of the data area displays various statistics about the named raid
    set. The statistics are - the current cache hit rate, the
    cumulative number of reads, read failures, writes and write
    failures for each backend of the raid set and finally the read and
    write throughput for each stargd process (indicated by it's process
    id) that front's the raid set.  The menu line allows one the leave
    the Statistics screen or select further actions.

    <Figure 11: Statistics Monitoring Screen>


 �  [Q]uit: Exit the Statistics screen and return to the General
    screen.

 �  [F]irst, [L]ast, [N]ext and [P]rev allow you to scroll through the
    statistics.

 �  [R]efresh: This option will get the statistics for the given raid
    set and re-display the current statistics on the screen.

 �  [Z]ero: This option will zero the cumulative statistics for the
    currently displayed raid set.

 �  [C]ontinuous: This option will start a back� ground process that
    will update the statis� tics of the currently displayed raid set
    every 2 seconds.  A loop counter is created and updated every 2
    seconds also. To inter� rupt this continuous mode of gathering
    statistics, just press any character. If you need to re-fresh the
    display, then press the refresh characters - <Control-l> or <Con�
    trol-r>.

 7.8.  [P]robe


 The probe option re-scans the SCSI channels and updates the backend
 list with the hardware it finds.

 7.9.  Example RAID Configuration Session


 The generalized procedure for configuration consists of three steps
 arranged in the following order:

 1. Configuring the Host Port(s)

 2. Assigning Spares

 3. Configuring the RAID set

 Note that there is a minimum number of backends required for the
 various supported RAID levels:

 �  Level 0 : 2 backends

 �  Level 3 : 2 backends

 �  Level 5 : 3 backends

 In this example we will configure a RAID 5 using 6, 2.04 gigabyte
 drives.  The total capacity of the virtual drive will be 10 gigabytes
 (the equivalent of one drive is used for redundancy). This same
 configuration procedure can be used to configure other levels of RAID
 sets by changing the type parameter.

 1. Power on the computer with the serial terminal connected to the
    RaidRunner's serial port.

 2. When the husky ( :raid; ) prompt appears, Start the GUI by typing
    "agui" and pressing return.

 3. When the main screen appears, select "H" for [H]ostport
    configuration

 4. On some models of RaidRunner the host port in not configurable. If
    you have only a [Q]uit option here then there is nothing further to
    be done for the host port configuration, note the values and skip
    to step 6. If you have add/modify options then your host port is
    software configurable.

 5. If there is no entry for a host port on this screen, add an entry
    with the parameters: controller=0, hostport=0 , SCSI ID=0. Don't
    forget to [I]nstall your changes. If there is already and entry
    present, note the values (they will be used in a later step).

 6. From this point onward I will assume the following hardware
    configuration:


    a. There are 7 - 2.04 gig drives connected as follows:

       i. 2 drives on SCSI channel 0 with SCSI IDs 0 and 1 (backends
          0.0.0, and 0.1.0, respectively).

       ii.
          3 drives on SCSI channel 1 with SCSI IDs 0 ,1 and 5 (backends
          1.0.0, 1.1.0, and 1.5.0).

       iii.
          2 drives on SCSI channel 2 with SCSI IDs 0 and 1 (backends
          2.0.0 and 2.1.0).

    b. Therefore:

       i. Rank 0 consists of backends 0.0.0, 1.0.0, 2.0.0

       ii.
          Rank 1 consists of backends 0.1.0, 1.1.0, 2.1.0

       iii.
          Rank 5 contains only the backend 1.5.0

    c. The RaidRunner is assigned to controller 0, hostport 0

 7. Press Q to [Q]uit the hostports screen and return to the Main
    screen.
 8. Press S to enter the [S]pares screen

 9. Select A to [A]dd a new spare to the spares pool. A list of
    available backends will be displayed and you will be prompted for
    the following information:


      Enter the device name to add to spares - from above:



 enter

 D1.5.0



 Select I to [I]nstall your changes

 Select Q to [Q]uit the spares screen and return to the Main screen

 Select R from the Main screen to enter the [R]aidsets screen.

 Select A to [A]dd a new RAID set. You will be prompted for each of the
 RAID set parameters. The prompts and responses are given below.

 1. Enter the name of Raid Set: cim_homes (or whatever you want to call
    it).

 2. Raid set type [0,1,3,5]: 5

 3. Enter initial host interface - ctlr,hostport,scsilun: 0.0.0

    Now a list of the available backends will be displayed in the form:
    0 - D0.0.0 1 - D1.0.0 2 - D2.0.0 3 - D0.1.0 4 - D1.1.0 5 - D2.1.0

 4. Enter index from above - Q to Quit: 1 press return 2 press return 3
    press return 4 press return 5 press return Q


 After pressing Q you will be returned to the Raid Sets screen. You
 should see the newly configured Raid set displayed in the data area.

 Press I to [I]nstall the changes

 <Figure 12: The RaidSets screen of the GUI showing the newly
 configured RAID 5>


 Press Q to exit the RaidSet screen and return to the the Main screen

 Press Q to [Q]uit agui and exit to the husky prompt.

 type "reboot" then press enter. This will reboot the RaidRunner (not
 the host machine.)

 When the RaidRunner reboots it will prepare the drives for the newly
 configured RAID.  NOTE: Depending on the size of the RAID this could
 take a few minutes to a few hours. For the above example it takes the
 5070 approximately 10 - 20 minutes to stripe the RAID set.

 Once you see the husky prompt again the RAID is ready for use. You can
 then proceed with the Linux configuration.

 8.  Linux Configuration


 These instructions cover setting up the virtual RAID drives on RedHat
 Linux 6.1. Setting it up under other Linux distributions should not be
 a problem.  The same general instructions apply.

 If you are new to Linux you may want to consider installing Linux from
 scratch since the RedHat installer will do most of the configuration
 work for you. If so skip to section titled "New Linux Installation."
 Otherwise go to the "Existing Linux Installation" section (next).

 8.1.  Existing Linux Installation


 Follow these instructions if you already have Redhat Linux installed
 on your system and you do not want to re-install. If you are
 installing the RAID as part of a new RedHat Linux installation (or are
 re-installing) skip to the "New Linux Installation" section.

 8.1.1.  QLogic SCSI Driver


 The driver can either be loaded as a module or compiled into your
 kernel.  If you want to boot from the RAID then you may want to use a
 kernel with compiled in QLogic support (see the kernel-HOWTO available
 from http://www.linuxdoc.org.  To use the modular driver become the
 superuser and add the following lines to /etc/conf.modules:

 alias qlogicpti /lib/modules/preferred/scsi/qlogicpti



 Change the above path to where ever your SCSI modules live. Then add
 the following line to you /etc/fstab (with the appropriate changes for
 device and mount point, see the fstab man page if you are unsure)

 /dev/sdc1 /home ext2 defaults 1 2



 Or, if you prefer to use a SYSV initialization script, create a file
 called "raid" in the /etc/rc.d/init.d directory with the following
 contents (NOTE: while there are a few good reasons to start the RAID
 using a script, one of the aforementioned methods would be
 preferable):

 #!/bin/bash



 case "$1" in



      start)



      echo "Loading raid module"


 /sbin/modprobe qlogicpti



 echo



 echo "Checking and Mounting raid volumes..."



 mount -t ext2 -o check /dev/sdc1 /home



 touch /var/lock/subsys/raid



 ;;



 stop)



      echo "Unmounting raid volumes"



 umount /home



 echo "Removing raid module(s)"



 /sbin/rmmod qlogicpti



 rm -f /var/lock/subsys/raid



 echo



 ;;



 restart)


      $0 stop



 $0 start



 ;;



 *)



      echo "Usage: raid {start|stop|restart}"



 exit 1



 esac



 exit 0



 You will need to edit this example and substitute your device name(s)
 in place of /dev/sdc1 and mount point(s) in place of /home. The next
 step is to make the script executable by root by doing:

 chmod 0700 /etc/rc.d/init.d/raid



 Now use your run level editor of choice (tksysv, ksysv, etc.) to add
 the script to the appropriate run level.

 8.1.2.  Device mappings


 Linux uses dynamic device mappings you can determine if the drives
 were found by typing:

 more /proc/scsi/scsi



 one or more of the entries should look something like this:

 Host: scsi1 Channel: 00 Id: 00 Lun: 00



 Vendor: ANTARES Model: CX106 Rev: 0109



 Type: Direct-Access ANSI SCSI revision: 02



 There may also be one which looks like this:

 Host: scsi1 Channel: 00 Id: 00 Lun: 07



 Vendor: ANTARES Model: CX106-SMON Rev: 0109



 Type: Direct-Access ANSI SCSI revision: 02



 This is the SCSI monitor communications channel which is currently un-
 used under Linux (see SMON in the advanced topics section below).

 To locate the drives (following reboot) type:

 dmesg | more



 Locate the section of the boot messages pertaining to you SCSI
 devices.  You should see something like this:

 qpti0: IRQ 53 SCSI ID 7 (Firmware v1.31.32)(Firmware 1.25 96/10/15)



 [Ultra Wide, using single ended interface]



 QPTI: Total of 1 PTI Qlogic/ISP hosts found, 1 actually in use.



 scsi1 : PTI Qlogic,ISP SBUS SCSI irq 53 regs at fd018000 PROM node ffd746e0



 Which indicates that the SCSI controller was properly recognized,
 Below this look for the disk section:

 Vendor ANTARES Model: CX106 Rev: 0109



 Type: Direct-Access ANSI SCSI revision: 02



 Detected scsi disk sdc at scsi1, channel 0, id 0, lun 0



 SCSI device sdc: hdwr sector= 512 bytes. Sectors= 20971200 [10239
  MB] [10.2 GB]



 Note the line that reads "Detected scsi disk sdc ..." this tells you
 that this virtual disk has been mapped to device /dev/sdc. Following
 partitioning the first partition will be /dev/sdc1, the second will be
 /dev/sdc2, etc. There should be one of the above disk sections for
 each virtual disk that was detected. There may also be an entry like
 the following:

 Vendor ANTARES Model: CX106-SMON Rev: 0109



 Type: Direct-Access ANSI SCSI revision: 02



 Detected scsi disk sdd at scsi1, channel 0, id 0, lun 7



 SCSI device sdd: hdwr sector= 512 bytes. Sectors= 20971200 [128 MB]
  [128.2 MB]



 BEWARE: this is not a drive DO NOT try to fdisk, mkfs, or mount it!!
 Doing so WILL hang your system.

 8.1.3.  Partitioning


 A virtual drive appears to the host operating system as a large but
 otherwise ordinary SCSI drive. Partitioning is performed using fdisk
 or your favorite utility. You will have to give the virtual drive a
 disk label when fdisk is started. Using the choice "Custom with
 autoprobed defaults" seems to work well. See the man page for the
 given utility for details.

 8.1.4.  Installing a filesystem


 Installing a filesystem is no different from any other SCSI drive:

 mkfs -t <filesystem_type> /dev/<device>



 for example:

 mkfs -t ext2 /dev/sdc1



 8.1.5.  Mounting


 If QLogic SCSI support is compiled into you kernel OR you are loading
 the "qlogicpti" module at boot from /etc/conf.modules then add the
 following line(s) to the /etc/fstab:

 /dev/<device> <mount point> ext2 defaults 1 1



 If you are using a SystemV initialization script to load/unload the
 module you must mount/unmount the drives there as well. See the
 example script above.

 8.2.  New Linux Installation


 This is the easiest way to install the RAID since the RedHat installer
 program will do most of the work for you.

 1. Configure the host port, RAID sets, and spares as outlined in
    "Onboard Configuration." Your computer must be on to perform this
    step since the 5070 is powered from the SBUS. It does not matter if
    the computer has an operating system installed at this point all we
    need is power to the controller card.

 2. Begin the RedHat SparcLinux installation

 3. The installation program will auto detect the 5070 controller and
    load the Qlogic driver

 4. Your virtual RAID drives will appear as ordinary SCSI hard drives
    to be partitioned and formatted during the installation. NOTE: When
    using the graphical partitioning utility during the RedHat
    installation DO NOT designate any partition on the virtual drives
    as type RAID since they are already hardware managed virtual RAID
    drives. The RAID selection on the partitioning utilities screen is
    for setting up a software RAID.  IMPORTANT NOTE: you may see a
    small SCSI drive ( usually ~128 MB) on the list of available
    drives. DO NOT select this drive for use. It is the SMON
    communication channel NOT a drive. If setup tries to use it it will
    hang the installer.

 5. Thats it, the installation program takes care of everything else !!

 9.  Maintenance


 9.1.  Activating a spare


 When running a RAID 3 or 5 (if you configured one or more drives to be
 spares) the 5070 will detect when a drive goes offline and
 automatically select a spare from the spares pool to replace it. The
 data will be rebuilt on-the-fly.  The RAID will continue operating
 normally during the re-construction process (i.e. it can be read from
 and written to just is if nothing has happened).  When a backend fails
 you will see messages similar to the following displayed on the 5070
 console:

 930 secs: Redo:1:1 Retry:1 (DIO_cim_homes_D1.1.0_q1) CDB=28(Read_10)Re-/Selection
  Time-out @682400+16



 932 secs: Redo:1:1 Retry:2 (DIO_cim_homes_D1.1.0_q1) CDB=28(Read_10)Re-/Selection
  Time-out @682400+16



 933 secs: Redo:1:1 Retry:3 (DIO_cim_homes_D1.1.0_q1) CDB=28(Read_10)Re-/Selection
  Time-out @682400+16



 934 secs: CIO_cim_homes_q3 R5_W(3412000, 16): Pre-Read drive 4 (D1.1.0)
  fails with result "Re-/Selection Time-out"



 934 secs: CIO_cim_homes_q2 R5: Drained alternate jobs for drive 4 (D1.1.0)



 934 secs: CIO_cim_homes_q2 R5: Drained alternate jobs for drive 4 (D1.1.0)
  RPT 1/0



 934 secs: CIO_cim_homes_q2 R5_W(524288, 16): Initial Pre-Read drive 4 (D1.1.0)
  fails with result "Re-/Selection Time-out"



 935 secs: Redo:1:0 Retry:1 (DIO_cim_homes_D1.0.0_q1) CDB=28(Read_10)SCSI
  Bus ~Reset detected @210544+16



 936 secs: Failed:1:1 Retry:0 (rconf) CDB=2A(Write_10)Re-/Selection Time-out
  @4194866+128



 Then you will see the spare being pulled from the spares pool, spun
 up, tested, engaged, and the data reconstructed.

 937 secs: autorepair pid=1149 /raid/cim_homes: Spinning up spare device



 938 secs: autorepair pid=1149 /raid/cim_homes: Testing spare device/dev/hd/1.5.0/data



 939 secs: autorepair pid=1149 /raid/cim_homes: engaging hot spare ...



 939 secs: autorepair pid=1149 /raid/cim_homes: reconstructing drive 4 ...



 939 secs: 1054



 939 secs: Rebuild on /raid/cim_homes/repair: Max buffer 2800 in 7491 reads,
  priority 6 sleep 500



 The rebuild script will printout its progress every 10% of the job
 completed


 939 secs: Rebuild on /raid/cim_homes/repair @ 0/7491



 1920 secs: Rebuild on /raid/cim_homes/repair @ 1498/7491



 2414 secs: Rebuild on /raid/cim_homes/repair @ 2247/7491



 2906 secs: Rebuild on /raid/cim_homes/repair @ 2996/7491



 9.2.  Re-integrating a repaired drive into the RAID (levels 3 and 5)


 After you have replaced the bad drive you must re-integrate it into
 the RAID set using the following procedure.

 1. Start the text GUI

 2. Look the list of backends for the RAID set(s).

 3. Backends that have been marked faulty will have a (-) to the right
    of their ID ( e.g. D1.1.0- ).

 4. If you set up spares the ID of the faulty backend will be followed
    by the ID of the spare that has replaced it ( e.g. D1.1.0-D1.5.0 )
    .

 5. Write down the ID(s) of the faulty backend(s) (NOT the spares).

 6. Press Q to exit agui

 7. At the husky prompt type:


      replace <name> <backend>



 Where <name> is whatever you named the raid set and <backend> is the
 ID of the backend that is being re-integrated into the RAID. If a
 spare was in use it will be automatically returned to the spares pool.
 Be patient, reconstruction can take a few minutes minutes to several
 hours depending on the RAID level and the size. Fortunately, you can
 use the RAID as you normally would during this process.


 10.  Troubleshooting / Error Messages


 10.1.  Out of band temperature detected...


 �  Probable Cause: The 5070 SBUS card is not adequately cooled.

 �  Solution: Try to improve cooling inside the case. Clean dust from
    the fans, re-organize the cards so the raid card is closest to the
    fan, etc. On some of the "pizza box" sun cases (e.g. SPARC 20) you
    may need to add supplementary cooling fans especially if you have
    it loaded with cards.
 10.2.  ... failed ... cannot have more than 1 faulty backend.


 �  Cause: More than one backend in the RAID 3/4/5 has failed (i.e.
    there is no longer sufficient redundancy to enable the lost data to
    be reconstructed).


 �  Solution: You're hosed ... Sorry.  If you did not assign spares
    when you configured you RAID 3/4/5 now may be a good time to re-
    consider the wisdom of that decision. Hopefully you have been
    making regular backups. Since now you will have to replace the
    defective drives, re-configure the RAID, and restore the data from
    a secondary source.

 10.3.  table.  When booting I see: ... Sun disklabel: bad magic 0000
 ... unknown partition


 �  Suspected Cause: Incorrect settings in the disk label set by fdisk
    (or whatever partitioning utility you used). This message seems to
    happen when you choose one of the preset disk labels rather than
    "Custom with autoprobed defaults."

 �  Solution: Since this error does not seem to effect the operation of
    the drive you can choose to do nothing and be ok. If you want to
    correct it you can try re-labeling the disk or re-partitioning the
    disk and choose "Custom with autoprobed defaults." If you are
    installing RedHat Linux from scratch the installer will get all of
    this right for you.

 11.  Bugs


 None yet! Please send bug reports to [email protected]

 12.  Frequently Asked Questions


 12.1.  How do I reset/erase the onboard configuration?


 At the husky prompt issue the following command:

 rconf -init



 This will delete all of the RAID configuration information but not the
 global variables and scsi monitors. the remove ALL configuration
 information type:

 rconf -fullinit



 Use these commands with caution!

 12.2.  How can I tell if a drive in my RAID has failed?


 In the text GUI faulty backends appear with a (-) to the right of
 their ID. For example the list of backends:

 D0.0.0,D1.0.0-,D2.0.0,D0.1.0,D1.1.0,D2.1.0

 Indicates that backend (drive) D1.0.0 is either faulty or not present.
 If you assigned spares (RAID 3 or 5) then you should also see that one
 or more spares are in use. Both the main and the and the RaidSets
 screens will show information on faulty/not present drives in a RAID
 set.

 13.  Advanced Topics: 5070 Command Reference


 In addition to the text based GUI the RAID configuration may also be
 manipulated from the husky prompt ( the : raid; prompt) of the onboard
 controller. This section describes commands that a user can input
 interactively or via a script file to the K9 kernel. Since K9 is an
 ANSI C Application Programming Interface (API) a shell is needed to
 interpret user input and form output. Only one shell is currently
 available and it is called husky. The K9 kernel is modelled on the
 Plan 9 operating system whose design is discussed in several papers
 from AT&T (See the "Further Reading" section for more information).
 K9 is a kernel targeted at embedded controllers of small to medium
 complexity (e.g. ISDN-ethernet bridges, RAID controllers, etc). It
 supports multiple lightweight processes (i.e. without memory
 management) on a single CPU with a non-pre-emptive scheduler. Device
 driver architecture is based on Plan 9 (and Unix SVR4) STREAMS.
 Concurrency control mechanisms include semaphores and signals. The
 husky shell is modelled on a scaled down Unix Bourne shell.

 Using the built-in commands the user can write new scripts thus
 extending the functionality of the 5070. The commands (adapted from
 the 5070 man pages) are extensive and are described below.

 13.1.  AUTOBOOT - script to automatically create all raid sets and
 scsi monitors


 �  SYNOPSIS: autoboot

 �  DESCRIPTION: autoboot is a husky script which is typically executed
    when a RaidRunner boots. The following steps are taken -

    1. Start all configured scsi monitor daemons (smon).

    2. Test to see if the total cache required by all the raid sets
       that are to boot is  not  more  than 90% of available memory.

    3. Start  all the scsi target daemons (stargd) and set each
       daemon's mode to "spinning-up" which enables it to respond to
       all non medium access commands from the host.  This  is  done
       to  allow hosts to gain knowledge about the RaidRunner's scsi
       targets as quickly as possible.

    4. Bind into the root (ram) filesystem all unused spare backend
       devices.

    5. Build all raid sets.

    6. If  battery  backed-up  ram is present, check for any saved
       writes and restore them into the just built raid sets.

    7. Finally, set the state of all scsi target daemons to "spun-up"
       enabling hosts to fully access the raid set's behind them.

 13.2.  failure AUTOFAULT - script to automatically mark a backend
 faulty after a drive



 �  SYNOPSIS: autofault raidset

 �  DESCRIPTION: autofault  is  a  husky  script  which is typically
    executed by a raid file system upon the failure of a backend of
    that raid set when that raid file system cannot use spare backends
    or has been configured not to use spare backends. After parsing
    it's arguments (command and environment) autofault issues a rconf
    command to mark a given backend as faulty.

 �  OPTIONS:

 �  raidset: The bind point of the raid set whose backend failed.

 �  $DRIVE_NUMBER: The index of the backend that failed. The first
    backend in a raid set is 0. This option is passed as an environment
    variable.

 �  $BLOCK_SIZE: The  raid set's io block size in bytes. (Ignored).
    This option is passed as an environment variable.

 �  $QUEUE_LENGTH: The raid set's queue length. (Ignored).  This option
    is passed as an environment variable.

 �  SEE ALSO: rconf

 13.3.  raid set AUTOREPAIR - script to automatically allocate a spare
 and reconstruct a


 �  SYNOPSIS: autorepair raidset size

 �  DESCRIPTION: autorepair  is  a  husky script which is typically
    executed by either a raid type 1, 3 or 5 file system upon the
    failure of a backend of that raid set.

    After parsing it's arguments (command and environment) autorepair
    gets a spare device from the  RaidRunner's  spares  spool.  It
    then  engages  it in write-only mode and reads the complete raid
    device which reconstructs the data on the spare.  The read is from
    the raid file system  repair  entrypoint.  Reading from  this
    entrypoint  causes  a  read  of  a  block immediately followed by a
    write of that block. The read/write sequence is atomic (i.e is not
    interruptible).  Once the reconstruction has completed, a check is
    made to ensure the spare did not fail during reconstruction and if
    not, the access mode of the spare device is set to the access mode
    of the raid set.  The process that  reads  the  repair  entrypoint
    is rebuild.

 This  device reconstruction will take anywhere from 10 minutes to one
 and a half hours depending on both the size and speed of the backends
 and the amount of activity the host is generating.

 During device reconstruction, pairs of numbers will be  printed
 indicating each  10%  of  data  reconstructed.  The pairs of numbers
 are separated by a slash character, the first number being the number
 of blocks reconstructed so far and the second being the number number
 of blocks to be reconstructed.  Further status about the rebuild can
 be gained from running rebuild.

 When the spare is allocated both the number of spares currently used
 on the backend and the spare device name is printed. The number of
 spares on a backend is referred to the depth of spares  on  the
 backend. Thus  prior to  re-engaging the spare after a reconstruction
 a check can be made to see if the depth is the same. If it is not,
 then the spare reconstruction failed and reconstruction using another
 spare  is underway (or no spares are available), and hence we don't
 re-engage the drive.


 �  OPTIONS:

 �  raidset: The bind point of the raid set whose backend failed.

 �  size : The size of the raid set in 512 byte blocks.

 �  $DRIVE_NUMBER: The  index  of  the  backend  that  failed. The
    first backend in a raid set is 0.  This option is passed as an
    environment variable.

 �  $BLOCK_SIZE: The raid set's io block size in bytes.  This option is
    passed as an environment variable.

 �  $QUEUE_LENGTH: The raid set's queue length.  This option is passed
    as an environment variable.

 �  SEE ALSO: rconf, rebuild

 13.4.  BIND - combine elements of the namespace


 �  SYNOPSIS: bind [-k] new old

 �  DESCRIPTION: Bind replaces the existing old file (or directory)
    with the new file (or directory). If the"-k" switch is given then
    new must be a kernel recognized device (file system). Section 7k of
    the manual pages documents the devices (sometimes called file
    systems) that can be bound using the "-k" switch.

 13.5.  BUZZER - get the state or turn on or off the buzzer


 �  SYNOPSIS: buzzer or buzzer on|off|mute

 �  DESCRIPTION: Buzzer will either print the state of the buzzer, turn
    on or off the buzzer or mute it. If no arguments  are  given then
    the state of the buzzer is printed, that is on or off will be
    printed if the buzzer is currently on or off respectively. If the
    buzzer has been muted, then you will be informed of this. If the
    buzzer has not been used since the RaidRunner has booted then the
    special state, unused, is printed. If the argument on is given the
    buzzer is turned on, if off, the buzzer is turned off. If the
    argument mute is given then the muted state of the buzzer is
    changed.

 �  SEE ALSO: warble, sos

 13.6.  CACHE - display information about and delete cache ranges


 �  SYNOPSIS: cache [-D moniker] [-I moniker] [-F] [-g moniker
    first|last] lastoffset

 �  DESCRIPTION: cache  will  print (to standard output) information
    about the given cache range, delete a given cache range, flush the
    cache or return the last offset of all cache ranges.

 �  OPTIONS

 �  -F: Flush all cache buffers to their backends (typically raid
    sets).


 �  -D moniker: Delete the cache range with moniker (name) moniker.

 �  -I moniker: Invalidate the cache for the given cache range
    (moniker). This is only useful for debugging or elaborate
    benchmarks.

 �  g moniker first|last: Print either the first or last block number
    of a cache range with moniker (name) moniker.

 �  lastoffset: Print the last offset of all cache ranges. The last
    offset is the last block number of all cache ranges.

 13.7.  CACHEDUMP - Dump the contents of the write cache to battery
 backed-up ram


 �  SYNOPSIS: cachedump

 �  DESCRIPTION: cachedump causes all unwritten data in the
    RaidRunner's cache to be written out to the battery backed-up ram.
    No data will be written to battery backed-up ram if there is
    currently valid  data  already  stored there.  This command is
    typically executed when there is something wrong with the data (or
    it's organization) in battery backed-up ram and you need to re-
    initialize it. cachedump will always return a NULL status.

 �  SEE ALSO: showbat, cacherestore

 13.8.  CACHERESTORE - Load the cache with data from battery backed-up
 ram


 �  SYNOPSIS: cacherestore

 �  DESCRIPTION: cacherestore will check the RaidRunner's battery
    backed-up ram for any data it has stored as a result of a power
    failure. It will copy any data directly into the cache. This
    command is typically executed automatically at boot time and prior
    to the  RaidRunner  making  it's data available to a host. Having
    successfully copied any data from battery backed-up ram into the
    cache, it flushes the cache and then re-initializes battery backed-
    up ram to indicate it holds no data. cacherestore will return a
    NULL status on success or 1 if an error occurred during the loading
    (with  a message written to standard error).

 �  SEE ALSO: showbat

 13.9.  CAT - concatenate files and print on the standard output


 �  SYNOPSIS: cat [ file... ]

 �  DESCRIPTION: cat writes the contents of each given file, or
    standard input if none are given or when a file named `-' is given,
    to standard output. If the nominated file is a directory then the
    filenames contained in that directory are sent to standard out (one
    per line). More information on a file (e.g. its size) can be
    obtained by using stat. The script file ls uses cat and stat to
    produce directory listings.

 �  SEE ALSO echo, ls, stat

 13.10.  CMP - compare the contents of 2 files



 �  SYNOPSIS: cmp [-b blockSize] [-c count] [-e] [-x] file1 file2

 �  DESCRIPTION: cmp compares the contents of the 2 named files. If
    file1 is "-" then standard input is used for that file. If the
    files are the same length and contain the same val� ues then
    nothing is written to standard output and the exit status NIL (i.e.
    true) is set. Where the 2 files dif� fer, the first bytes that
    differ and the position are out� put to standard out and the exit
    status is set to "differ" (i.e. false). The position is given by a
    block number (origin 0) followed by a byte offset within that block
    (origin 0). The optional "-b" switch allows the blockSize of each
    read operation to be set. The default blockSize is 512 (bytes). For
    big compares involving disks a relatively large blockSize may be
    useful (e.g. 64k). See suffix for allowable suffixes. The optional
    "-c" switch allows the count of blocks read to fixed. A value of 0
    for count is interpreted as read to the end of file (EOF). To
    compare the first 64 Megabytes of 2 files the switches "-b 64k -c
    1k" could be used. See suffix for allowable suffixes. The optional
    "-e" switch instructs ccmmpp to output to stan� dard out (usually
    overwriting the same line) the count of blocks compared, each time
    a multiple of 100 is reached. The final block count is also output.
    The optional "-x" switch instructs ccmmpp to continue after a
    comparison error (but not a file error) and keep a count of blocks
    in error. If any errors are detected only the last one will be
    output when the command exits. If the "-e" switch is also given
    then the current count of blocks in error is output to the right of
    the multiple of 100 blocks compared. This command is designed to
    compare very large files. Two buffers of blockSize are allocated
    dynamically so their size is bounded by the amount of memory (i.e.
    RAM in the target) available at the time of command execution. The
    count could be up to 2G. The number of bytes compared is the
    product of blockSize and count (i.e.  big enough).

 �  SEE ALSO: suffix

 13.11.  CONS - console device for Husky


 �  SYNOPSIS: bind -k cons bind_point

 �  DESCRIPTION: cons  allows  an interpreter (e.g. Husky) to route
    console input and output to an appropriate device. That console
    input and output is available at bind_point in the K9 namespace.
    The special file cons should always be available.

 �  EXAMPLES: Husky does the following in its initialisation:


      bind -k cons /dev/cons



 On a Unix system this is equivalent to:

 bind -k unixfd /dev/cons



 On a DOS system this is equivalent to:

 bind -k doscon /dev/cons



 On target hardware using a SCN2681 chip this is equivalent to:

 bind -k scn2681 /dev/cons



 SEE ALSO: unixfd, doscon, scn2681

 13.12.  DD - copy a file (disk, etc)


 �  SYNOPSIS: dd [if=file] [of=file] [ibs=bytes] [obs=bytes] [bs=bytes]
    [skip=blocks] [seek=blocks] [count=blocks] [flags=verbose]

 �  DESCRIPTION: dd copies a file (from the standard input to the
    standard output, by default) with a user-selectable blocksize.

 �  OPTIONS

 �  if=file Read from file instead of the standard input.

 �  of=file, Write to file instead of the standard output.

 �  ibs=bytes, Read given number of bytes at a time.

 �  obs=bytes, Write given number of bytes at a time.

 �  bs=bytes, Read and write given number of bytes at a time. Override
    ibs and obs.

 �  skip=blocks, Skip ibs-sized blocks at start of input.

 �  seek=blocks, By-pass obs-sized blocks at start of output.

 �  count=blocks, Copy only ibs-sized input blocks.

 �  flags=verbose, Print (to standard output) the number of blocks
    copied every ten percent of the copy. The output is of the form X/T
    where X is the number of blocks copied so far and T is the total
    number of blocks to copy. This option can only be used if both the
    count= and of= options  are also given.


    The decimal numbers given to "ibs", "obs", "bs", "skip", "seek" and
    "count" must not be negative. These numbers can optionally have a
    suffix (see suffix). dd outputs to standard out in all cases. A
    successful copy of 8 (full) blocks would cause the following
    output:

    8+0 records in



 8+0 records out



 The number after the "+" is the number of fractional blocks (i.e.
 blocks that are less than the block size) involved. This number will
 usually be zero (and is otherwise when physical media with alignment
 requirements is involved).

 A write failure outputting the last block on the previous example
 would cause the following output:

 Write failed



 8+0 records in



 7+0 records out



 SEE ALSO: suffix

 13.13.  DEVSCMP - Compare a file's size against a given value


 �  SYNOPSIS: devscmp filename size

 �  DESCRIPTION: devscmp will find the size of the given file and
    compare it's size in 512-byte blocks to the given size (to be in
    512-byte blocks).  If the size of the file is less than the given
    value, then -1 is printed, if equal to then 0 is printed, and if
    the size of the given file is greater than the given size then 1 is
    printed. This routine is used in internal scripts to ensure that
    backends of raid sets are of an appropriate size.

 13.14.  DFORMAT- Perform formatting functions on a backend disk drive


 �  SYNOPSIS

 �  dformat -p c.s.l -R bnum

 �  dformat -p c.s.l -pdA|-pdP|-pdG

 �  dformat -p c.s.l -S [-v] [-B firstbn]

 �  dformat -p c.s.l -F

 �  dformat -p c.s.l -D file

 �  DESCRIPTION: In it's first form dformat will either reassign a
    block on a nominated disk drive.  via the SCSI-2 REASSIGN BLOCKS
    command. The second form will allow you to print out the current
    manufacturers  defect  list  (-pdP), the  grown defect  list (-pdG)
    or both defect lists (-pdA). Each printed list is sorted with one
    defect per line in Physical Sector Format - Cylinder Number, Head
    Number and Defect Sector Number. The third form causes the drive to
    be scanned in a destructive write/read/compare manner. If a  read
    or write  or data comparison error occurs then an attempt is made
    to identify the bad sector(s).  Typically the drive is scanned from
    block 0 to the last block on the drive.  You can optionally give an
    alternative starting block number. The fourth form causes a low
    level format on the specified device. The fifth option allows you
    to download a device's microcode into the device.

 �  OPTIONS:

 �  -R bnum: Specify a logical block number to reassign to the drive's
    grown defect list.

 �  -pdA: Print both the manufacturer's and grown defect list.


 �  \ -pdP: Print the manufacturer's defect list.

 �  -pdG: Print the grown defect list.

 �  -S: Perform a destructive scan of the disk reporting I/O errors.

 �  -B firstbn: Specify the first logical block number to start a scan
    from.

 �  -v: Turn on verbose mode - which prints the current block number
    being scanned.

 �  -F: Issue a low-level SCSI format command to the given device. This
    will take some time.

 �  -D file: Download  into  the  specified device, the given file. The
    download is effected by a single SCSI Write-Buffer command in save
    microcode mode. This allows users to update  a  device's
    microcode. Use this command carefully as you could destroy the
    device by loading an incorrect file.

 �  -p c.s.l: Identify  the disk device by specifying it's channel,
    SCSI ID (rank) and SCSI LUN provided in the format "c.s.l"

 �  SEE ALSO: Product manual for disk drives used in your RAID.

 13.15.  DIAGS - script to run a diagnostic on a given device


 �  SYNOPSIS: diags disk -C count -L length -M io-mode -T io-type -D
    device

 �  DESCRIPTION: diags is a husky script which is used to run the
    randio diagnostic on a given device. When randio is executed, it is
    executed in verbose mode.

 �  OPTIONS:

 �  disk: This is the device type of diagnostic we are to run.

 �  -C count: Specify the number of times to execute the diagnostic.

 �  -L length: Specify  the  "length" of the diagnostic to execute.
    This can be either short, medium or long and specified with the
    letter's s, m or l respectively. In the case of a disk, a short
    test will  the first 10% of the device, a medium the first 50% and
    long the whole (100%) of the disk.

 �  -M io-mode: Specify a destructive (read-write) or non-destructive
    (read-only) test.  Use either read-write or read-only.

 �  -T io-type: Specify a type of io - either sequential or random.

 �  -D device: Specify the device to test.

 �  SEE ALSO: randio, scsihdfs

 13.16.  DPART - edit a scsihd disk partition table


 �  SYNOPSIS:

 �  dpart -a|d|l|m -D file [-N name] [-F firstblock] [-L lastblock]

 �  dpart -a -D file -N name -F firstblock -L lastblock

 �  dpart -d -D file -N name

 �  dpart -l -D file

 �  dpart -m -D file -N name -F firstblock -L lastblock

 �  DESCRIPTION: Each  scsihd device  (typically a SCSI disk drive) can
    be divided up into eight logical partitions. By default when a
    scsihd device is bound into the RaidRunner's file system, it has
    four  partitions, the  whole device  (raw),  typically named
    bindpoint/raw, the partition file (bindpoint/partition), the
    RaidRunner backup configuration file (bindpoint/rconfig), and the
    "data" portion of the disk (bind- point/data) which represents the
    whole device less the backup configuration area and partition file.
    For more information, see scsihdfs. If other partitions are added,
    then they will appear as bindpoint/partitionname.  dpart allows you
    to edit or list the partition table on a scsihd device (typically a
    disk).

 �  OPTIONS:

 �  -a: Add a partition. When adding a partition, you need to specify
    the partition name  (-N)  and  the partition range from the first
    block (-F) to the last block (-L).

 �  -d: Delete a named (-N) partition.

 �  -l: List all partitions.

 �  -m: Modify an existing partition.  You  will need to specify the
    partition name (-N) and BOTH it's first (-F) and last (-L)
    blocknumbers even if you are just modifying the last block number.

 �  -D file: Specify the partition file to be edited. Typically, this
    is the bindpoint/partition file.

 �  -N name: Specify the partition name.

 �  -F firstblock: Specify the first block number of the partition.

 �  -L lastblock: Specify the last block number of the partition.

 �  SEE ALSO: scsihd

 13.17.  DUP - open file descriptor device


 �  SYNOPSIS: bind -k dup bind_point

 �  DESCRIPTION: The dup device makes a one level directory with an
    entry in that directory for every open file descriptor of the
    invoking K9 process. These directory "entries" are the numbers.
    Thus a typical process (script) binding a dup device would at least
    make these files in the namespace: "bind_point/0", "bind_point/1"
    and "bind_point/2". These would correspond to its open standard in,
    standard out and standard error file descriptors.  A dup device
    allows other K9 processes to access the open file descriptors of
    the invoking process. To  do  this  the  other processes  simply
    "open" the  required  dup  device directory entry whose name (a
    number) corresponds to the required file descriptor.

 13.18.  ECHO - display a line of text


 �  SYNOPSIS: echo [string ...]

 �  DESCRIPTION: echo writes each given string to the standard output,
    with a space between them and a newline after the last one. Note
    that all the string arguments are written in a single write kernel
    call. The following backslash-escaped characters in the strings are
    converted as follows:

    \b     backspace

 \c     suppress trailing newline

 \f     form feed

 \n     new line

 \r     carriage return

 \t     horizontal tab

 \v     vertical tab

 \\     backslash

 \nnn   the character whose ASCII code is nnn (octal)


 �  SEE ALSO: cat

 13.19.  ENV- environment variables file system


 �  SYNOPSIS: bind -k env bind_point

 �  DESCRIPTION: env file system associates a one level directory with
    the bind_point in the K9 namespace. Each file name in that
    directory is the name of the environment variable while the
    contents of the file is that variable's current value. Conceptually
    each process sees their own copy of the env file system.  This copy
    is either empty or inherited from this process's parent at spawn
    time (depending on the flags to spawn).

 13.20.  ENVIRON - RaidRunner Global environment variables - names and
 effects


 �  DESCRIPTION: The RaidRunner uses GLOBAL environment variables to
    control the functionality of automatic actions. GLOBAL environment
    variables are saved in the Raid configuration area so they retain
    their values between reboots/power downs. Certain RaidRunner
    internal run-time variables can also be set as a GLOBAL environment
    variables. See the internals manual entry for details. The table
    below describes those GLOBAL environment variables that are used by
    the RaidRunner in it's normal operation.

 �  RebuildPri

    This  variable,  if  set, controls the priority used when drive
    reconstruction occurs via the rebuild program. If the variable is
    not set then the default rebuild priority would be used. The
    variable is to be a comma separated list of raid set names and
    their  associated  rebuild  priorities  and  sleep periods (colon
    separated). The form is

    Rname_1:Pri_1:Sleep_1,Rname_2:Pri_2:Sleep_2,...,Rname_N:Pri_N:Sleep_N



 where  Pri_1  is  to  be  the priority the rebuild program runs with
 when run on raid set Rname_1, Sleep_1 is the period, in milliseconds,
 to sleep between each rebuild action on the raid set, Pri_2 is to be
 the priority  for  raid set Rname_2, and so forth.  For example, if
 the value of RebuildPri is

 R:5:30000



 then if a rebuild occurs (via replace, repair or autorepair) on raid
 set R then the rebuild will run with priority 5 (via the -p rebuild
 option) and will sleep 30000 milliseconds (30 seconds) between each
 rebuild action (specified via the -S rebuild option). The priority
 given must be valid for the rebuild program.

 �  BackendRanks

    On  certain  RaidRunner's where multiple controllers may exist, you
    can restrict a controller's access to the backend ranks of devices
    available.  For example, you may have 2 controllers and 4 ranks of
    backend devices. You can  specify that  the  first controller can
    only access the first two ranks and the second controller, the
    second two ranks. This variable along with other associated
    commands allows you to set up this restriction. Additionally, you
    may only have a single controller RaidRunner which is in  an
    enclosure  with multiple  ranks.  By default  the controller will
    attempt to probe for all devices on all ranks. If you have only
    populated the RaidRunner with say, half it's possible compliment of
    backend devices, then the RaidRunner will still probe for the other
    half. Setting this variable appropriately will prevent this un-
    needed (and on occasion time consuming) process. This variable
    takes the form

    controller_id:ranklist controller_id:ranklist ...



 where controller_id is the controller number (from 0 upwards) and
 ranklist is a comma list of backend ranks which the given controller
 will access. Note that the backend rank is the scsi-id of that rank.
 For example, on a 2 rank (rank 1 and 2 - i.e scsi id 1 for the first
 rank and scsi id 2 for the second), 1  controller


 This variable takes the form


 For example, on a 2 rank (rank 1 and 2 - i.e scsi id 1 for the first
 rank and scsi id 2 for the second), 1  controller RaidRunner  where
 only  the first  rank  has devices you could prevent the controller
 from attempting to access the (empty) second rank by setting Back�
 endRanks to

 0:1



 Typically, you would not set this variable directly, but use support�
 ing commands  to  set  it.  These  commands  are pranks and sranks.
 See these manual entries for details.


 �  RAIDn_reference_PBUFS

    Raid  types 3, 4 and 5 all make use of memory for temporary parity
    buffers when they need to create parity data. This memory is in
    addition to that allocated to a raid set's cache.  When a raid set
    is created, it  will  also  create a default  number  of  parity
    buffers (which are the same size is a raid set's iosize).
    Sometimes, if the iosize of the raid set is large there will not be
    enough memory to create this default number of parity buffers.  To
    overcome this situation, you can set GLOBAL environment variables
    to over-ride the default number of parity buffers that all raid
    sets of a particular type or a specific raid set will use. You need
    to set these variables before you define the raid set  via  agui
    and if you delete them and not the raid set, then the effect raid
    sets may not boot and hence will not be accessible by a host. The
    variables are of the form RAIDn_reference_PBUFS where n is the raid
    type (3, 4 or 5), and reference  is the raid set's name or the
    string 'Default' You use the reference of 'Default' to specify all
    raid sets of a particular type.  For example, to over-ride the
    number of parity buffers for a raid 5 named

    : raid ; setenv RAID5_FRED_PBUFS 64



 To over-ride the number of parity buffers for ALL raid 3's (and set
 only 72 parity buffers) set

 : raid ; setenv RAID3_Default_PBUFS 128



 If you set a default for all raid sets of a particular type, but want
 ONE of them to be different then set up a  variable  for that
 particular  raid set as it's value will over-ride the default. In the
 above example, where all Raid Type 3 will have 128 parity buffers, you
 could set the variable

 : raid ; setenv RAID3_Dbase_PBUFS 56



 which will allow the raid 3 raid set named 'Dbase' to have 56 parity
 buffers, but all other raid 3's defined on the RaidRunner will have
 128.


 �  SEE ALSO: setenv, printenv, rconf, rebuild, internals

 13.21.  EXEC - cause arguments to be executed in place of this shell


 �  SYNOPSIS: exec [ arg ... ]

 �  DESCRIPTION: exec causes the command specified by the first arg to
    be executed in place of this shell without creating a new process.
    Subsequent args are passed to the command specified by the first
    arg as its arguments. Shell redirection may appear and, if no other
    arguments are given, causes the shell input/output to be modified.

 13.22.  EXIT - exit a K9 process


 �  SYNOPSIS: exit [string]


 �  DESCRIPTION: exit has an optional string argument. If the optional
    argument is given the current K9 process is terminated with the
    given string as its exit value. (If the string has embedded spaces
    then the whole string should be a quoted_string). If no argument is
    given then the shell gets the string associated with the
    environment variable "status" and returns that string as the exit
    value. If the environment variable "status" is not found then the
    "true" exit status (i.e. NIL) is returned.

 �  SEE ALSO: true, K9exit

 13.23.  EXPR - evaluation of numeric expressions


 �  SYNOPSIS: expr numeric_expr ...

 �  DESCRIPTION: expr evaluates each numeric_expr command line argument
    as a separate numeric expression. Thus a single expression cannot
    contain unescaped whitespaces or needs to be placed in a quoted
    string (i.e.  between "{" and "}"). Arithmetic is performed on
    signed integers (currently numbers in the range from -2,147,483,648
    to 2,147,483,647). Successful calculations cause no output (to
    either standard out/error or environment variables). So each useful
    numeric_expr needs to include an assignment (or op-assignment).
    Each numeric_expr argument supplied is evaluated in the order given
    (i.e. left to right) until they all evaluate successfully
    (returning a true status). If evaluating a numeric_expr fails
    (usually due to a syntax error) then the expr command fails with
    "error" as the exit status and the error message is written to the
    environment variable "error".

 �  OPERATORS: The precedence of each operator is shown following the
    description in square brackets. "0" is the highest precedence.
    Within a single precedence group evaluation is left-to-right except
    for assignment operators which are right-to-left. Parentheses have
    higher precedence than all operators and can be used to change the
    default precedence shown below.

    UNARY OPERATORS

 +

 Does nothing to expression/number to the right.

 -

 negates expression/number to the right.

 !

 logically negate expression/number to the right.

 ~

 Bitwise negate expression/number to the right.

 BINARY ARITHMETIC OPERATORS

 *

 Multiply enclosing expressions [2]

 /

 Integer division of enclosing expressions

 %

 Modulus of enclosing expressions.

 +

 Add enclosing expressions

 -

 Subtract enclosing expressions.

 <<

 Shift left expression _left_ by number in right expression. Equivalent
 to: left * (2 ** right)

 >>

 Shift left expression _right_ by number in right expression.
 Equivalent to: left / (2 ** right)

 &

 Bitwise AND of enclosing expressions

 ^

 Bitwise exclusive OR of enclosing expressions. [8]

 |

 Bitwise OR of enclosing expressions. [9]

 BINARY LOGICAL OPERATORS

 These logical operators yield the number 1 for a true comparison and 0
 for a false comparison. For logical ANDs and ORs their left and right
 expressions are assumed  to be false if 0 otherwise true. Both logical
 ANDs and ORs evaluate both their left and right expressions in all
 case (cf. C's short-circuit action).

 <=

 true when left less than or equal to right. [5]

 >=

 true when left greater than or equal to right. [5]

 <

 true when left less than right. [5]

 >

 true when left greater than right. [5]

 ==

 true when left equal to right. [6]

 !=

 true when left not equal to right. [6]

 &&

 logical AND of enclosing expressions [10]

 ||

 logical OR of enclosing expressions [11]

 ASSIGNMENT OPERATORS

 In the following descriptions "n" is an environment variable while
 "r_exp" is an expression to the right. All assignment operators have
 the same precedence which is lower than all other operators. N.B.
 Multiple assignment operators group right-to-left (i.e. same as C
 language).

 =

 Assign right expression into environment variable on left.

 *=

 n *= r_exp is equivalent to: n = n * r_exp

 /=

 n /= r_exp is equivalent to: n = n / r_exp

 %=

 n %= r_exp is equivalent to: n = n % r_exp

 +=

 n += r_exp is equivalent to: n = n + r_exp

 -=

 n -= r_exp is equivalent to: n = n - r_exp

 <<=

 n <<= r_exp is equivalent to: n = n << r_exp

 >>=

 n >>= r_exp is equivalent to: n = n >> r_exp

 &=

 n &= r_exp is equivalent to: n = n & r_exp

 |=

 n |= r_exp is equivalent to: n = n | r_exp


 �  NUMBERS: All number are signed integers in the range stated in the
    description above. Numbers can be input in base 2 through to base
    36. Base 10 is the default base. The default base can be overridden
    by:

    1. a leading "0" : implies octal or hexadecimal

    2. a number of the form _base_#_num_

    Numbers prefixed with "0" are interpreted as octal. Numbers pre�
    fixed with "0x" or "0X" are interpreted as hexadecimal.  For  num�
    bers  using the "#" notation the _base_ must be in the range 2
    through to 36 inclusive. For bases greater then 10 the letters "a"
    through "z" are utilised for the extra "digits".  Upper and lower
    case letters are acceptable. Any single digit that exceeds (or is
    equal to) the base is consider an error. Base 10 numbers only may
    have a suffix. See suffix for a list of valid suffixes. Also note
    that since expr uses signed integers then "1G" is the largest mag�
    nitude number that can be represented with the "Gigabyte" suffix
    (assuming 32 bit signed integers, -2G is invalid due to the order
    of evaluation).


 �  VARIABLES: The only symbolic variables allowed are K9 environment
    variables.  Regardless of whether they are being read or written
    they should never appear preceded by a "$". Environment variables
    that didn't previous exist that appear as left argument of an
    assignment are created. When a non-existent environment variable is
    read then it is interpreted as the value 0.

 �  EXAMPLES: Some simple examples:


      expr {n = 1 + 2} # create n



 echo $n



 3



 expr {n*=2} # 3 * 2 result back into n



 echo $n



 6



 expr { k = n > 5 } # 6 > 5 is true so create k = 1



 echo $k



 1



 �  NOTE: expr is a Husky "built-in" command. See the "Note" section in
    "set" to see the implications.


 �  SEE ALSO: husky, set, suffix, test


 13.24.  FALSE - returns the K9 false status


 �  SYNOPSIS: false

 �  DESCRIPTION: false does nothing other than return a K9 false
    status. K9 processes return a pointer to a C string (null
    terminated array of characters) on termination. If that pointer is
    NULL then a true exit value is assumed while all other returned
    pointer values are interpreted  as false (with the string being
    some explanation of what went wrong). This command returns a
    pointer to the string "false" as its return value.

 �  EXAMPLE: The following script fragment will print "got here" to
    standard out:


      if false then



 echo impossible



 else



 echo got here



 end



 �  SEE ALSO: true


 13.25.  FIFO - bi-directional fifo buffer of fixed size


 �  SYNOPSIS:

 �  bind -k {fifo size} bind_point

 �  cat bind_point

 �  bind_point/data

 �  bind_point/ctl

 �  DESCRIPTION: fifo  file system associates a one level directory
    with the bind_point in the K9 namespace with a buffer size of size
    bytes. bind_point/data and bind_point/ctl are the data and control
    channels for the fifo. Data written to the bind_point/data file is
    available for reading from the same file in a first-in first-out
    basis. A write of x bytes to the bind_point/data file will either
    complete and and transfer all the data, or will  transfer
    sufficient bytes until the fifo buffer is full then block until
    data is removed from the fifo buffer by reading. A read of x bytes
    from the bind_point/data file will transfer the lessor of the
    current amount of data in the fifo buffer or x bytes.  A read from
    the bind_point/ctl will return the size of the fifo buffer and the
    current usage. The number of opens (#  Opens) is the number of
    processes that currently have the bind_point/data file open.

 �  EXAMPLE


      > /buffer



 bind -k {fifo 2048} /buffer



 ls -l /buffer



 /buffer:



 /buffer/ctl                     fifo    2 0x00000001    1 0



 /buffer/data                    fifo    2 0x00000002    1 0



 cat /buffer/ctl



 Max: 2048 Cur: 0, # Opens: 0



 echo hello > /buffer/data



 cat /buffer/ctl



 Max: 2048 Cur: 6, # Opens: 0



 dd if=/buffer/data bs=512 count=1



 hello



 0+1 records in

 0+1 records out



 cat /buffer/ctl



 Max: 2048 Cur: 0, # Opens: 0



 �  SEE ALSO: pipe

 13.26.  GET - select one value from list


 �  SYNOPSIS: get number [ value ... ]

 �  DESCRIPTION: get uses the given number to select one value from the
    given list. Indexing is origin 0 (e.g. "get 0 aaa bb c" returns
    "aaa").  If the number is out of range for an index on the given
    list of values then nothing is returned.

 13.27.  GETIV - get the value an internal RaidRunner variable


 �  SYNOPSIS:

 �  getiv

 �  getiv name

 �  DESCRIPTION: getiv prints the current value of an internal
    RaidRunner variable or prints a list of all variables. When a
    variable name is given it's current value is printed. If no value
    is given the all available internal variables are listed.

 �  NOTES: As different models of RaidRunners have different internal
    variables see your RaidRunner's Hardware Reference manual for a
    list of variables together with the meaning of their values. These
    variables are run-time variables and hence revert to their default
    value whenever  the  RaidRunner is booted.

 �  SEE ALSO: setiv

 13.28.  HELP - print a list of commands and their synopses


 �  SYNOPSIS: help or ?

 �  DESCRIPTION: help or the question mark character - ?, will print a
    list of all commands available to the command interpreter. Along
    with each command, it's synopsis is printed.

 13.29.  HUSKY - shell for K9 kernel


 �  SYNOPSIS

 �  husky [-c command] [ file [ arg ... ] ]

 �  hs [-c command] [ file [ arg ... ] ]

 �  DESCRIPTION: husky and hs are synonyms. husky is a command language
    interpreter that executes commands read from the standard input or
    from a file. husky is a scaled down model of Unix's Bourne shell
    (sh). One major difference is that husky has no concept of current
    working directory. If the "-c" switch is present then the following
    command is interpreted by husky in a newly thrown shell nested in
    the current environment. This newly thrown shell exits back to the
    current environment when the command finishes. Otherwise if
    arguments are given the first one is assumed to be a file
    containing husky commands.  Again a new shell is thrown to execute
    these commands. husky script files can access their command line
    arguments and the 2nd and subsequent arguments to husky (if
    present) are passed to the file for that purpose. If no arguments
    are given to husky then commands are read from standard in (and the
    shell is considered interactive).

 �  RETURN STATUS: husky places the K9 return status of a process (NIL
    if ok, otherwise a string explaining the error) in the file
    "/env/status"

    An example:

    dd if=/xx



 dd: could not open /xx



 cat /env/status



 open failed



 cat /env/status



 # empty because previous "cat" worked



 As the file "/env/status" is an environment variable the return status
 of a command is also available in the variable $status. The exit
 status of a pipeline is the exit status of the last command in the
 pipeline.


 �  SIGNALS If an interactive shell receives an interrupt signal (i.e.
    K9_SIGINT - usually a control-C on the console) then the shell
    exits. The "init" process will then start a new instance of the
    husky shell with all the previously running processes (with the
    exception of the just killed shell) still running.  This allows the
    user to kill the process that caused the previous shell problems.
    Alternatively a process that is acci� dentally run in foreground is
    effectively put in the background by sending an interrupt signal to
    the shell. Note that this is quite different to Unix shells which
    would forward the signal onto the foreground process.

 �  QUOTES, ESCAPING, STRING CONCATENATION, ETC: A quoted_string (as
    defined in the grammar) commences with a "{" and finishes with the
    matching "}". The term "matching"  implies  that all embedded "{"
    must have a corresponding embedded "}" before the final "}" is said
    to match the original "{".  A quoted_string can be spread across
    several lines. No command line substitution occurs within
    quoted_strings. The character for escaping the following character
    is "\". If a "{" needs to be interpreted literally then it can be
    represented by "\{". If a string containing spaces (whitespaces)
    needs to be interpreted as a single token then space (whitespace)
    can be escaped (i.e. "\ ").  If a "\" itself needs to be
    interpreted literally then it can be represented by "\\".  The
    string concatenation character is "^". This is useful when a token
    such as "/d4" needs to built up by a script when "/d" is  fixed
    and  the "4" is derived from some variable:


      set n 4



 > /d^$n



 This example would create the file "/d4".

 The output of another husky command or script can be made available
 inline by starting the sequence with "`" and finishing it with a "'".
 For example:

 echo {ps output follows:



 } `ps'



 This prints the string "ps output follows:" followed on the next line
 by the current output from the command "ps". That output from  "ps"
 would have its embedded newlines replaced by whitespaces.


 COMMAND LINE FILE REDIRECTION:

 �  Redirection should appear after a command and its arguments in a
    line to be interpreted by husky. A special case is a line that just
    contains "> filename" which creates the filename with zero length
    if it didn't previously exist or truncates to zero length if it
    did.

 �  Redirection of standard in to come from a file uses the token "<"
    with the filename appearing to its right. The default source of
    standard in is the console.

 �  Redirection of standard out to go to a file uses the token ">" with
    the filename appearing to its right. The default destination of
    standard out is the console.

 �  Redirection of standard error to go to a file uses the token ">[2]"
    with the filename appearing to its right. The default destination
    of standard error is the console.

 �  Redirection of writes from within a command which uses a known file
    descriptor number (say "n") to go to a file uses the token ">[n]"
    with the filename appearing to its right.


 �  Redirection of read from within a command which uses a known file
    descriptor number (say "n") to come from a file uses the token
    "<[n]" with the filename appearing to its right.

 �  Redirection of reads and writes from within a command which uses a
    known file descriptor number (say "n") to a file uses the token
    "<>[n]" with the filename appearing to its right. In order to
    redirect both standard out and standard error to the one file the
    form " > filename >[2=1]" can be used. This sequence first
    redirects standard  out  (i.e. file descriptor 1) to filename and
    then redirects what is written to file descriptor 2 (i.e.  standard
    error) to file descriptor 1 which is now associated with filename.


 ENVIRONMENT VARIABLES: Each process can access the name it was invoked
 by via the variable: "arg0" . The command line arguments (excluding
 the invocation name) can be accessed as a list in the variable: "argv"
 . The number of elements in the list "argv" is place in "argc".  The
 get command is useful for fetching individual arguments from this
 list.  The pid of the current process can be fetched from the
 variable: "pid".  When a script launches a new process in the
 background then the child's pid can be accessed from the variable
 "child". The variable "ContollerId" is set to the RaidRunner
 controller number husky is running on. Environment variables are a
 separate "space" for each process. Depending on the way a process was
 created, its initial set of environment variables may be copied from
 its parent process at the "spawn" point.

 SEE ALSO: intro

 13.30.  HWCONF - print various hardware configuration details


 �  SYNOPSIS: hwconf [-D] [-M] [-I] [-d [-n]] [-f] [-h] [-i -p c.s.l]
    [-m] [-p c.s.l] [-s] [-S] [-t] [-T] [-P] [-W]

 �  DESCRIPTION: hwconf prints details about the RaidRunner hardware
    and devices attached.

 �  OPTIONS:

 �  -h: Print  the number of controllers, host interfaces per
    controller, the number of disk channels per controller, number of
    ranks of disks and the details memory (in bytes) on each
    controller.  Four memory  figures  are printed, the first is the
    total memory in the controller, next is the amount of memory at
    boot time, next is the amount currently available and lastly is the
    largest available contiguous area of memory.  This is the default
    option.

 �  -f: Print  the  number  of  fans in the RaidRunner and then the
    speed for each fan in the system. The speeds values are in
    revolutions per minute (rpms).  The fans in the system are  labeled
    in  your hardware  specification sheet for your RaidRunner. The
    first speed printed from this command corresponds to fan number 0
    on your specification sheet, the second is for fan 1, and so forth.

 �  -d: Print out information on all the disk drives on the RaidRunner.
    For each disk on the RaidRunner, print out - the  device  name,  in
    the format c.s.l where c is the channel, s is the SCSI ID (or rank)
    and l is the SCSI LUN of the device, the manufacturer's name
    (vendor id), the disk's model name (product id), the disk's version
    id, the disk serial number, the disk geometry - number of
    cylinders, heads and sectors, and the last block number on the disk
    and the block size in bytes. the disk revolution count per minute
    (rpm's), the number of notches/zones available on the drive (if
    any)

 �  -n: Print out the disk drive notch/zone tables if available.  This
    is a sub-option to the -d  option. Not  all  disks appear to
    correctly report the notch/zone partition tables.  For each
    notch/zone,

 �  the following is printed: the zone number, the zone's starting
    cylinder, the zone's starting head, the zone's ending cylinder, the
    zone's ending head, the zone's starting logical block number, the
    zone's ending logical block number, the zone's number of sectors
    per track

 �  -D: Print out the device names for all disk drives on the system.

 �  -I: Initialize back-end NCR SCSI chips.  This flag may be used in
    conjunction with any  other  option and  will  done first. It has
    an effect only the first call to hwconf that has not yet used a -d,
    -D or -I options,  or on those chips that have not yet had a -p on
    the  channel  associated  with that chip.

 �  -m: Print  out  major  flash and battery backed-up ram addresses
    (in hex).  Additionally print out the size of the RaidRunner
    configuration area.  Eight (8) addresses are printed in order
    RaidRunner configuration area start and end addresses (FLASH RAM),
    RaidRunner Husky Scripts area start and end addresses (FLASH RAM),
    RaidRunner Binary Image area start and end addresses (FLASH RAM),
    RaidRunner Battery Backed-up area start and end addresses. And the
    size of the RaidRunner configuration area (in bytes) is then
    printed.

 �  -p c.s.l: Probe a single device specified by the given channel,
    SCSI ID (rank) and SCSI LUN provided in the format  "c.s.l". The
    output of this command is the same as the "-d" option but just for
    the given device. If the device is not present then nothing will be
    output and the exit status of the  command will be 1.

 �  -i -p c.s.l: Re-initialize  the SCSI device driver specified by the
    given channel, SCSI ID (rank) and SCSI LUN provided in the format
    "c.s.l".  Typically this command is used when, on a running
    RaidRunner,  a new drive is plugged in, and it will be used prior
    to the RaidRunner's next reboot.

 �  -M: Set  the  boottime  memory. This option is executed internally
    by the controller at boot time and has no function (or effect)
    executed at any other time.

 �  -s: Print the 12 character serial number of the RaidRunner.

 �  -S: Issue SCSI spin up commands to all backends as quickly as
    possible.  This option is  intended  for use at power-on stage
    only.

 �  -t: Probe  the  temperature  monitor  returning the internal
    temperature of the RaidRunner in degrees Celsius.

 �  -T: Print the temperatures being recorded by the hardware
    monitoring daemon (hwmon).

 �  -P: For both AC and DC power supplies, print the number of each
    present and the state of each supply. The state will be printed as
    ok or flt depending on whether the PSU is working or faulty.

 �  -W: This  option  will  wait until all possible backends have spun
    up.  It is used in conjunction with

 �  NOTES : The order of printing the disk information is by SCSI ID
    (rank), by channel, by SCSI LUN.

 13.31.  HWMON - monitoring daemon for temperature, fans, PSUs.


 �  SYNOPSIS: hwmon [-t seconds] [-d]

 �  DESCRIPTION: hwmon  is  a  hardware  monitoring  daemon.  It
    periodically probes the status of certain elements of a RaidRunner
    and if an out-of-band occurrence happens, will cause the alarm to
    sound  or  light  up  fault leds as well as saving a message in the
    system log. Depending on the model of RaidRunner, the elements
    monitored are temperature, fans and power supplies. When an out-of-
    band occurrence is found, hwmon will reduce the time between probes
    to 5 seconds.  If  a  buzzer is the alarm device, then the buzzer
    will turn on for 5 seconds then off for 5 seconds and repeat this
    cycle until the buzzer is muted or the occurrence is corrected.

    If the RaidRunner model supports a buzzer muting switch, then the
    buzzer will be muted if the switch  is pressed  during  a  cycle
    change as per the previous paragraph. When hwmon recognizes the
    mute switch it will beep twice.

 Certain out-of-band occurrences can be considered to be catastrophic,
 meaning if the occurrence  remains uncorrected,  the  RaidRunner's
 hardware is likely to be damaged. Occurrences such as total fan
 failure and sustained high temperature along with total or partial fan
 failure are considered  as  catastrophic. hwmon has  a  means  of
 automatically placing the RaidRunner into a "shutdown" or quiescent
 state where minimal power is consumed (and hence less heat is
 generated). This is  done  by  the  execution  of  the shutdown
 command  after a period of time where catastrophic out-of-band
 occurrences are sustained.  This process is enabled, via the
 AutoShutdownSecs internal variable. See the internals  manual  for use
 of this variable. hwmon  can  be  prevented from starting at boot time
 by creating the global environment variable NoHwmon and setting any
 value to it. A warning message will be stored in the syslog.


 �  OPTIONS:

 �  t seconds: Specify the number of seconds to wait between probes of
    the hardware elements.  If this option is not specified, the
    default period is 300 seconds.

 �  -d: Turn on debugging mode which can produce debugging output.

 �  SEE ALSO: hwconf, pstatus, syslogd, shutdown, internals

 13.32.  running kernel INTERNALS - Internal variables used by RaidRun�
 ner to change dynamics of


 �  DESCRIPTION: Certain  run-time  features  of the RaidRunner can be
    manipulated by changing internal variables via the setiv command.
    The table below describes each changeable variable, it's effect,
    it's default value and range of values it can be set to. The
    variables  below  are  run-time  features  of a RaidRunner and
    hence are always set to their default values when a RaidRunner
    boots. Certain variables can be stored as a global environment
    variable and will over-ride the  defaults  at boot time.  If you
    create a global environment variable of that variable's name with
    an appropriate value, it's default value will be over-ridden the
    next time the RaidRunner is re-booted. Note, that the values of
    these variables  ARE  NOT CHECKED  when set in the global
    environment variable tables and, if incorrectly set, will generate
    errors at boot until deleted or corrected.  In the table below, any
    variable that can have a value stored as a global  environment
    variable is marked with (GEnv)

 �  write_limit: This  variable  is  the  maximum  number of 512-byte
    blocks the cache filesystem will buffer for writes. If this limit
    is reached all writes to the cache filesystem will be blocked until
    the cache filesystem has  written out (to it's backend) enough
    blocks to reach a low water mark - write_low_tide.  This  variable
    cannot  be  changed if battery backed-up RAM is available as it is
    tied to the amount of battery backed-up RAM available. The value of
    this variable is calculated when the cache is initialized. It's
    value is dependant on whether  battery  backed-up RAM is installed
    in the RaidRunner. If installed, the number of blocks of data that
    can be saved into the battery backed-up RAM is calculated. If no
    battery backed-up RAM is present, it's value is set  to 75% of  the
    RaidRunner's  memory  (expressed  in a count of 512 byte blocks)
    then adjusted to reflect the amount of cache requested by
    configured raid sets. When write_limit is changed then both
    write_high_tide and write_low_tide are  automatically  changed  to
    there default values (a function of the value of write_limit).

 �  write_high_tide: This variable is a high water mark for the number
    of written-to 512-byte blocks in the cache. When the number of data
    blocks exceeds this value, to avoid the cache filesystem from
    blocking it's front end, the  cache  flushing mechanism
    continually  flushes the cache buffer until the amount of unwritten
    (to the backend) cache buffers is below the low water mark
    (write_low_tide).  This value defaults to 75% of write_limit. This
    variable can have values ranging from write_limit down to
    write_low_tide. It is recommended that this variable not be
    changed.

 �  write_low_tide: This variable is a low water mark for when the
    cache flushing mechanism is continually  flushing  data  to  it's
    backend. Once  the  number of written-to cache blocks yet to be
    flushed equals or is less than this value, the sustained flushing
    is stopped. This value defaults to 25% of write_limit.  This
    variable can have values ranging from write_high_tide-1 down to
    zero (0).  It is recommended that this variable not be changed.

 �  cache_nflush: This variable is the number of cache buffers (not
    512-byte data blocks) that the cache flushing  mechanism  will
    attempt to write out in one flush cycle. Adjusting  this  value
    may improve performance on writes depending of the size of the
    cache buffers and type of disk drives used in the raid set
    backends. The default value is 128. It's value can range from 2 to
    128.

 �  cache_nread: This variable is the number of cache buffers (not
    512-byte data blocks) that the cache  reading  mechanism  will
    attempt to read out in one read cycle. Adjusting  this  value  may
    improve performance on reads depending of the size of the cache
    buffers and type of disk drives used in the raid set backends. The
    default value is 128. It's value can range from 2 to 128.

 �  cache_wlimit: This variable is the number of cache buffers (not
    512-byte data blocks) that the cache flushing  mechanism  will
    attempt  coalesce  into a  single sequential write. It is different
    to cache_nflush in that cache_nflush is the total number of cache
    buffers that can be written in a single cache flush cycle and these
    buffers  can  be  non sequential  whereas  cache_wlimit is a limit
    on the number of sequential cache buffer's that can be written with
    one write. Adjusting this value may improve performance on writes
    depending of the size of the cache buffers  and  type  of disk
    drives used in the raid set backends. The default value is 128.
    It's value can range from 2 to 128.

 �  cache_fperiod (GEnv): By  default, the cache flushes any data to be
    written every 1000 milliseconds (unless it's forced to by the fact
    that the cache is getting full and then it flushes the cache and
    resets the timer).  You can vary this  flushing period  by setting
    this variable.  Given you have a large number of sustained reads
    and minimal writes, then you may want to delay the writes out of
    cache to the backends as long as possible. Note, that by setting
    this  to  a high value, you run the risk of loosing what you have
    written.  The default value is 1000 milliseconds (i.e 1 second).
    It's value can range from 500ms to 300000ms.

 �  scsi_write_thru (GEnv): By  default all writes (from a host) are
    buffered in the RaidRunner's cache and are flushed to the backend
    disks periodically.  When battery backed-up RAM is available then
    this results in the most efficient write  throughput. If no battery
    backed-up RAM is available or you do not want to depend on writes
    being saved in battery backed-up RAM in event of a power failure
    you can force the RaidRunner to write data straight thru to the
    backends prior to returning an OK status to the host. This
    essentially provides a write-thru cache. The default value of this
    variable is 0 - write-thru mode is DISABLED.  The values this
    variable can take are

 �  0 - DISABLE write-thru mode, or

 �  1 - ENABLE write-thru mode.

 �  scsi_write_fua (GEnv): This variable effects what is done when the
    FUA (Force Unit Access) bit is set on a SCSI WRITE-10 command. When
    this  variable is  enabled  and a SCSI WRITE-10 command has the FUA
    bit set is processed then the data is written directly thru the
    cache to the backend disks. If the variable is disabled, then the
    setting of  the  FUA bit on SCSI WRITE-10 commands is ignored. The
    default value for this variable is disabled (0) if battery backed-
    up RAM is present, or enabled (1) if battery backed-up RAM is NOT
    present.  The values this variable can take are

 �  0 - IGNORE FUA bit on SCSI WRITE-10 commands, or

 �  1 - ACT on FUA bit on SCSI WRITE-10 commands.

 �  scsi_ierror (GEnv): This variable controls what is done when the
    RaidRunner receives a Initiator Detected Error message  on  a  SCSI
    host channel. If set  (1), cause an Check Condition, If NOT set
    (0), follow the SCSI-2 standard and re-transmit the Data In / Out
    phase. The default value is 0. The values this variable can take
    are

 �  0 - follow SCSI-2 standard

 �  1 - ignore the SCSI-2 standard and cause a Check Condition.

 �  scsi_sol_reboot (GEnv): Determines whether to auto-detect a Solaris
    reboot and the clear any wide mode negotiations. If set (1), detect
    a Solaris reboot and clear wide mode.  If NOT set (0), follow the
    SCSI-2  standard  and  not clear wide mode. The default value is 0.
    The values this variable can take are

 �  0 - follow SCSI-2 standard

 �  1 - ignore the SCSI-2 standard and clear wide mode.


 �  scsi_hreset (GEnv): Determines whether to issue a SCSI bus reset on
    host ports after power-on. If  set  (1), then a SCSI bus reset is
    done on the host port when starting the first smon/stargd process
    on that port.  If NOT set (0), nothing is done. The default value
    is 0. The values this variable can take are

 �  0 - don't issue SCSI bus resets on power-on.

 �  1 - issue SCSI bus resets on power-on when the first smon/stargd
    process is started.

 �  scsi_full_log (GEnv): Determines whether or not stargd reports, via
    syslog, a Reset Check condition on Read, Write,  Test  Unit  Ready
    and Start Stop commands.  This reset check condition is always set
    when a RaidRunner boots or the raid detects a scsi-bus reset. Note
    that this variable only suppresses the logging of this Check
    condition into syslog, it does not effect the response to the host
    of this and any Check condition. If  set  (1),  then  all stargd
    detected reset Check condition error messages are logged.  If NOT
    set (0), these messages are suppressed The default value is 0. The
    values this variable can take are

 �  0 - suppress logging these messages

 �  1 - log all messages.

 �  scsi_ms_badpage (GEnv): Determines whether or not stargd reports,
    via syslog, that it has received a non-supported page number in a
    MODE SENSE  or  MODE SELECT command it receives from a host. Note
    that stargd will issue the appropriate Check condition to the host
    ("Invalid Field in CDB") irrespective of the value of this
    variable. If set (1), then all stargd detected non-supported page
    numbers in MODE SENSE and MODE SELECT commands  will  be   logged.
    If NOT set (0), these messages are suppressed The default value is
    0. The values this variable can take are

 �  0 - suppress logging these messages

 �  1 - log all messages.

 �  scsi_bechng (GEnv): Determines  whether  or not the raid reports
    backend device parameter change errors. In a multi controller
    environment, backends are probed and some of their parameters are
    changed by a booting controller.  This will  generate parameter
    change mode sense errors. If cleared (0), then all parameter change
    errors will NOT be logged.  If set (1), these messages are logged
    like any other backend error. The default value is 0. The values
    this variable can take are

 �  0 - suppress logging these messages

 �  1 - log all messages.

 �  scsi_dnotch (GEnv): Some disk drives take an inordinate amount of
    time to perform mode select commands.  One  set  of  information  a
    RaidRunner will  obtain from a device backend are the disk notch
    pages (if present). As this is for information only, then to reduce
    the boot time of a RaidRunner you can request that disk notches are
    not obtained. If cleared (0), backend disk notch information is not
    probed for.  If set (1), then backend disk notch information is
    probed for. The default value is 1. The values this variable can
    take are:

 �  0 - don't probe for notch pages


 �  1 - probe for notch pages

 �  scsi_rw_retries (GEnv): Specify  the  number  of  read  or write
    retries to perform on a device backend before effecting an error on
    the given operation.  Note that ALL retries are reported via
    syslog. The default value is 3. It's value can range from 1 to 9.

 �  scsi_errpage_r (GEnv): Specify the number of internal read retries
    that a disk backend is to perform before reporting an error (to
    the raid). Setting this variable causes the Read Retry Count field
    in the Read-Write Error Recovery mode sense page. A value of -1
    will cause the drive's default to be used. The default value is -1.
    It's value can range from -1 (use disk's default) or from 0 to 255.

 �  scsi_errpage_w (GEnv): Specify the number of internal write retries
    that a disk backend is to perform before reporting an error (to the
    raid).  Setting this  variable  causes  the Write Retry Count field
    in the Read-Write Error Recovery mode sense page.  A value of -1
    will cause the drive's default to be used. The default value is -1.
    It's value can range from -1 (use disk's default) or from 0 to 255.

 �  BackFrank: Specify the SCSI-ID of the first rank of backend disks
    on a RaidRunner.  This variable should never  be  changed and is
    for informative purposes only. The default value is dependant on
    the model of RaidRunner being run. The values this variable can
    take are

 �  0 - the first rank SCSI-ID will be 0

 �  1 - the first rank SCSI-ID will be 1

 �  raid_drainwait (GEnv): Specify  the  number  of  milliseconds  a
    raidset is to delay, before draining all backend I/O's when a
    backend fails. Setting this variable to a lower value will speed up
    the commencement of any  error recovery  procedures that would be
    performed on a raid set when a backend fails. The default value is
    500 milliseconds. It's value can range from 50 to 10000
    milliseconds.

 �  EnclosureType: Specify  the  enclosure  type a raid controller is
    running within.  This variable should never be changed and is for
    informative purposes only. The default value is dependant on the
    model of RaidRunner being run.  The values this variable can take
    are integers starting from 0.

 �  fmt_idisc_tmo (GEnv): Specify the SCSI command timeout (in
    milliseconds) when a SCSI FORMAT command  is  issued  on  a
    backend.  Disk drives take different amounts of time to perform a
    SCSI FORMAT command and hence a timeout is required to be set when
    the command is issued. As certain drives may take longer to format
    than the default timeout you can  change it. The default value is
    720000 milliseconds. It's value can range from 200000 to 1440000
    milliseconds.

 �  AutoShutdownSecs (GEnv): Specify  the  number  of seconds the
    RaidRunner should monitor catastrophic hardware failures before
    deciding to automatically shutdown. A catastrophic failure is one
    which will cause damage to the  RaidRunner's hardware  if not
    fixed  immediately.  Failures  like all fans failing would be
    considered catastrophic. A value of 0 seconds (the default) will
    disable this feature, that is, with the exception of  logging  the
    errors,  no  action will occur.  See the shutdown and hwmon for
    further details. The default value is 0 seconds. It's value can
    range from 20 to 125 seconds.


 �  SEE ALSO: setiv, getiv, syslog, setenv, printenv, hwmon, shutdown

 13.33.  KILL - send a signal to the nominated process


 �  SYNOPSIS: kill [-sig_name] pid

 �  DESCRIPTION: kill sends a signal to the process nominated by pid.
    If the pid is a positive number then only the nominated process is
    signaled. If the pid is a negative number then the signal is sent
    to all processes in the same process group as the process with the
    id of -pid. The switch is optional and if not given a SIGTERM
    (software termination signal) is sent. If the sig_name switch is
    given then it should be one of the following (lower case)
    abbreviations.  Only the first 3 letters need to be given for the
    signal name to be recognized.  Following each abbreviation is a
    brief explanation and the signal number in brackets:

    null - unused signal [0]

 hup - hangup [1]

 int - interrupt (rubout) [2]

 quit - quit (ASCII FS) [3]

 kill - kill (cannot be caught or ignored) [4]

 pipe - write on a pipe with no one to read it [5]

 alrm - alarm clock [6]

 term - software termination signal [7]

 cld - child process has changed state [8]

 nomem - could not obtain memory (from heap) [9]

 You cannot kill processes whose process id is between 0 and 5
 inclusive.  These are considered sacrosanct - hyena, init and console
 reader/writers.


 �  SEE ALSO: K9kill

 13.34.  LED- turn on/off LED's on RaidRunner


 �  SYNOPSIS:

 �  led

 �  led led_id led_function

 �  DESCRIPTION: led uses the given led_id to identify the LED to
    manipulate based on the led_function. When no arguments are given,
    an internal LED register is printed along with the current function
    the onboard LEDS, led1 and led2 are tracing. If a undefined led_id
    is given, the led command silently does nothing and returns NULL.
    If an incorrect number of arguments or invalid led_function is
    given a usage message is printed. Depending on the RaidRunner model
    the led_id can be one of

 �  led1 - LED1 on the RaidRunner controller itself


 �  led2 - LED2 on the RaidRunner controller itself

 �  Dc.s.l - Device on channel c, scsi id s, scsi lun l

 �  status - the status LED on the RaidRunner

 �  io - the io LED on the RaidRunner


    and led_function can be one of

 �  on - turn on the given LED

 �  off - turn off the given LED

 �  ok - set the given LED to the defined OK state

 �  faulty - set the given LED to the defined FAULTY state

 �  warning - set the given LED to the defined WARNING state

 �  rebuild - set the given LED to the defined REBUILD state

 �  tprocsw - set the given LED to trace kernel process switching

 �  tparity - set the given LED to trace I/O parity generation

 �  tdisconn - set the given LED to trace host interface disconnect
    activity

 �  pid - set the given LED to trace the process pid as it runs

 Different models of RaidRunner have various differences in number of
 LED's and their functionality.  Depending on the type of LED, the ok,
 faulty, warning and rebuild functions perform different functions. See
 your RaidRunner's Hardware Reference manual to see what LED's exist
 and what different functions do.


 NOTES: Tracing activities can only occur on the `onboard` leds (LED1,
 LED2).

 SEE ALSO: lflash

 13.35.  LFLASH- flash a led on RaidRunner


 �  SYNOPSIS: lflash led_id period

 �  DESCRIPTION: lflash uses the given led_id to identify the LED to
    flash every period seconds. If a undefined led_id is given, the led
    command silently does nothing and returns NULL. Depending on the
    RaidRunner model the led_id can be one of:

    led1 - LED1 on the RaidRunner controller itself

 led2 - LED2 on the RaidRunner controller itself

 Dc.s.l - Device on channel c, scsi id s, scsi lun l

 status - the status LED on the RaidRunner

 io - the io LED on the RaidRunner



 �  NOTE: The number of seconds must be greater than or equal to 2.

 �  SEE ALSO: led

 13.36.  LINE - copies one line of standard input to standard output


 �  SYNOPSIS: line

 �  DESCRIPTION: line accomplishes the one line copy by reading up to a
    newline character followed by a single K9write.

 �  SEE ALSO: K9read, K9write

 13.37.  LLENGTH - return the number of elements in the given list


 �  SYNOPSIS: llength list

 �  DESCRIPTION: llength returns the number of elements in a given
    list.

 �  EXAMPLES: Some simple examples:


      set list D1 D2 D3 D4 D5 # create the list



 set len `llength $list' # get it's length



 echo $len



 5



 set list {D1 D2 D3 D4 D5} {D6 D7}  # create the
  list



 set len `llength $list' # get it's length



 echo $len



 2



 set list {} # create an empty list



 set len `llength $list' # get it's length

 echo $len



 0



 13.38.  LOG - like zero with additional logging of accesses


 �  SYNOPSIS: bind -k {log fd error_rate tag} bind_point

 �  DESCRIPTION: log is a special  file that when written to is a
    infinite sink of data (i.e. anything can be written to it and it
    will be disposed of quickly). When log is read it is an infinite
    source of zeros (i.e. the byte value 0). The log file will appear
    in the K9 namespace at the bind_point. Additionally, ASCII  log
    data  is  written  to  the file associated with file descriptor fd.
    error_rate should be a number between 0 and 100 and is the
    percentages of errors (randomly distributed) that will be reported
    (as an  EIO  error) to  the caller. Each line written to fd will
    have tag appended to it. There is one line output to fd for each IO
    operation on the log special file. The first character output is
    "R" or "W" indicating a read or write. The second character is
    blank if no error was reported and "*" if one was reported.  Next
    (after  a white  space)  is  a (64 bit integer) offset into the
    file of the start of the operation, followed by the size (in bytes)
    of that operation. The line finishes with the tag.

 �  EXAMPLE: Bind a log special file at "/dev/log" that writes log
    information to standard error. Each line written to standard error
    has the tag string "scsi" appended to it.  Approximately 30% of
    reads and writes (i.e. randomly distributed) return an EIO error to
    the caller. This is done as follows:


      bind "log 2 30 scsi" /dev/log



 dd if=/dev/zero of=/dev/log count=5 bs=512



 W  0000000000 512        scsi



 W  0000000200 512        scsi



 W  0000000400 512        scsi



 W* 0000000600 512        scsi



 Write failed.

 4+0 records in



 3+0 records out



 SEE ALSO: zero

 13.39.  LRANGE - extract a range of elements from the given list


 �  SYNOPSIS: lrange first last list

 �  DESCRIPTION: lrange returns a list consisting of elements first
    through last of list. 0 refers to the first element in the list. If
    first is greateR THAN last then the list is extracted in reverse
    order.

 �  EXAMPLES: Some simple examples:

    set list D1 D2 D3 D4 D5 # create the list



 set subl `lrange 0 3 $list' # extract from indices 0 to 3



 echo $subl



 D1 D2 D3 D4



 set subl `lrange 3 1 $list' # extract from indices 3 to 1



 echo $subl



 D4 D3 D2



 set subl `lrange 4 4 $list' # extract from indices 0 to 3



 echo $subl # equivalent to get 4 $list



 D5



 set subl `lrange 3 100 $list'


 echo $subl



 D4 D5



 13.40.  LS - list the files in a directory


 �  SYNOPSIS: ls [ -l ] [ directory... ]

 �  DESCRIPTION: ls lists the files in the given directory on standard
    out.  If no directory is given then the root directory (i.e. "/")
    is listed. Each file name contained in a directory is put on a
    separate line.  Each listing has a lead-in line stating which
    directory is being shown. If there  is  more than one directory
    then they are listed sequentially separated by a blank line. If the
    "-l" switch is given then every listed file has data such as its
    length and the file system it belongs to shown on the same line as
    its name. See the stat command for more information. ls is not an
    inbuilt command but a husky script which utilizes cat and stat. The
    script source can be found in the file "/bin/ps".

 �  SEE ALSO: cat, stat

 13.41.  LSEARCH - find the a pattern in a list


 �  SYNOPSIS: lsearch pattern list

 �  DESCRIPTION: lsearch returns the index of the first element in list
    that matches pattern or -1 if none. 0 refers to the first element
    in the list

 �  EXAMPLES: Some simple examples:

    set list D1 D2 D3 D4 D5 # create the list



 set idx `lsearch D4 $list' # get index of D4 in list



 echo $idx



 3



 set idx `lsearch D1 $list' # get index of D1 in list



 echo $idx



 0



 set idx `lsearch D8 $list' # get index of D8 in list



 echo $idx # equivalent to get 4 $list



 -1



 13.42.  LSUBSTR - replace a character in all elements of a list


 �  SYNOPSIS: lsubstr find_char replacement_char list

 �  DESCRIPTION: lsubstr returns a list replacing every find_ch
    character found in any element of the list with the
    replacement_char character. replacement_char can be NULL  which
    effectively deletes all find_char characters in the list.

 �  EXAMPLES: Some simple examples:

    set list D1 D2 D3 D4 D5 # create the list



 set subl `lsubstr D x $list' # replace all D's with x's



 echo $subl



 x1 x2 x3 x4 x5



 set subl `lsubstr D {} $list' # delete all D's



 echo $subl



 1 2 3 4 5



 set list -L -16 # create a list with embedded braces



 set subl `lsubstr {} $list' # delete all open braces



 set subl `lsubstr {} $subl' # delete all close braces



 echo $subl

 -L 16



 13.43.  MEM - memory mapped file (system)


 �  SYNOPSIS: bind -k {mem first last [ r ]} bind_point

 �  DESCRIPTION: mem  allows  machine memory to be accessed as a single
    K9 file (rather than a file system). The host system's memory is
    used starting at the first memory location up to and including the
    last memory location.  Both first and last need to be given  in
    hexadecimal. If successful the mem file will appear in the K9
    namespace at the bind_point. The stat command will show it as a
    normal file with the appropriate size (i.e. last - first + 1).  If
    the optional "r" is given then only read-only access to the file is
    permitted. In a target environment mem can usefully associate
    battery backed-up RAM (or ROM) with the K9 namespace. In a Unix
    environment  it  is of  limited  use (see unixfd instead). In a DOS
    environment it may be useful to access memory directly (IO space)
    but for accessing the DOS console see doscon. When mem is
    associated with the partition of Flash RAM that stores the husky
    scripts, which is  stored  compressed, reading from that page will
    automatically decompress and return the data as it is read. When
    mem  is associated with the writable partitions of Flash RAM
    (configuration partition, husky script partition and main binary
    partition) a write to the start of any partition will erase that
    partition.

 �  SEE ALSO: ram

 �  BUGS: Only a single file rather than a file system can be bound.

 13.44.  MDEBUG - exercise and display statistics about memory alloca�
 tion


 �  SYNOPSIS: mdebug [off|on|trace|p|m size|f ptr|c nel elsize|r ptr
    size]

 �  DESCRIPTION: mdebug  can  be  used to directly allocate and free
    memory.  mdebug will also print (to standard output) information
    about the current state of memory allocation.  With out any given
    options a brief five  line summary of memory usage is printed, e.g.


      : raid; mdebug



 Mdebug is off



 nreq-nfree=87096-82951=4145(13905745)



 size=15956672/16150000



 waste=1%/2%



 list=4251/8396



 : raid;



 The first line indicates the debug mode,  either off, on or trace.
 The second line indicates the number times a request for memory is
 made (to Mmalloc() or Mcalloc() and related functions) and the  number
 of times  the  memory allocator is called to free memory (via
 Mfree()).  The difference between these first two numbers is the total
 number of currently allocated blocks of memory, with the  number
 between  the '('  and  ')'  being  the total memory requested.  Note
 that the amount of memory actually assign may be more than requested.

 The third line indicates the amount of memory being managed.  The
 second number is the total memory man aged (i.e. left over after
 loading the statically allocated text, data and bss space).  The first
 number is that left over after various memory allocation tables have
 been subtracted out from that  afore mention  number.  The fourth line
 is the total amount of extra memory assigned to requests in excess of
 the actual requested memory as compared with the totals on line 3.

 The fifth line relates to the list of currently allocated memory.  The
 first number  is  the  number  of free entries left and the second is
 the maximum table size.  Note that the number of currently allocated
 blocks (third number on line 2) when added to the first number on line
 5 gives the second number on line 5.


 OPTIONS:

 �  p: Prints the above mentioned five line summary and then the free
    list.

 �  P: Prints all the above plus dumps the list of currently allocated
    memory.

 �  PP: Prints all the above plus the free bitmap.

 The  above  three  options can generate copious output and require a
 detailed knowledge of the source to understand their meaning.

 off: Turns off memory allocation debugging.  This is the default
 condition after booting.

 on: Turns on memory allocation assertion checking.

 trace: Turns on memory allocation assertion checking and traces every
 memory allocation /  deallocation.

 m: Uses Mmalloc() to allocate a block of memory of size bytes.

 f: Uses Mfree() to de-allocate a block of memory addressed by ptr.

 c: Uses  Mcalloc()  to allocate a contiguous block of memory
 consisting of nel elements each of size bytes.

 r: Uses Mrealloc() to re-allocate a block of previously allocated
 memory, ptr,  changing  the  allocated size to be size bytes.
 SEE ALSO: Unix man pages on malloc()

 13.45.  MKDIR - create directory (or directories)


 �  SYNOPSIS: mkdir [ directory_name ... ]

 �  DESCRIPTION: mkdir creates the given directory (or directories). If
    all the given directories can be created then NIL is returned as
    the status; otherwise the first directory that could not be created
    is returned (and this command will continue trying to create
    directories until the list is exhausted). A directory cannot be
    created with a file name that previously existed in the enclosing
    directory.

 13.46.  MKDISKFS - script to create a disk filesystem


 �  SYNOPSIS: mkdiskfs disk_directory_root disk_name

 �  DESCRIPTION: mkdiskfs  is  a  husky  script which is used to
    perform all the necessary commands to create a disk filesystem
    given the root of the disk file system and the name of the disk.

 �  OPTIONS :

 �  disk_directory_root: Specify the directory root under which the
    disk filesystems are bound.  This is typically /dev/hd.

 �  disk_name: Specify the name of the disk in the format Dc.s.l where
    c is the channel, s is the scsi  id  (or  rank) and l is the scsi
    lun of the disk.



    After parsing it's arguments mkdiskfs creates the disk filesystem's
    bind point and binds in the disk at that point. set.


 SEE ALSO: rconf, scsihdfs

 13.47.  MKHOSTFS - script to create a host port filesystem


 �  SYNOPSIS: mkhostfs controller_number host_port host_bus_directory

 �  DESCRIPTION: mkhostfs  is a husky script which is used to perform
    all the necessary commands to create a host port filesystem on the
    given RaidRunner controller given the root of the host port file
    systems and the host port number.

 �  OPTIONS:

 �  controller_number: Specify the controller on which the host port
    filesystem is to be created.

 �  host_port: Specify the host port number to create the filesystem
    for.

 �  host_bus_directory: Specify the directory root under which host
    filesystems are bound. This is typically /dev/hostbus. After
    parsing  it's  arguments mkhostfs finds out what SCSI ID the host
    port is to present (see hconf and then binds in the host
    filesystem. set.


 �  SEE ALSO: hconf, scsihpfs

 13.48.  MKRAID - script to create a raid given a line of output of
 rconf


 �  SYNOPSIS: mkraid `rconf -list RaidSetName'

 �  DESCRIPTION: mkraid  is a husky script which is used to perform all
    the necessary commands to create and enable host access to a given
    Raid Set. The arguments to mkraid is a line of output from a rconf
    -list  command. After parsing  it's  arguments mkraid  checks  to
    see if a reconstruction was being performed when the RaidRunner was
    last operating, and if so, notes this.  It then creates the raid
    filesystem (see mkraidfs) and adds a cache frontend to the  raid
    filesystem.  It then creates the required host filesystems (see
    mkhsotfs) and finally, if a reconstruction had been taking place
    when the RaidRunner was last operating, it restarts a
    reconstruction.

 �  NOTE: This husky script DOES NOT enable target access (stargd) to
    the raid set it creates.

 �  SEE ALSO: rconf, mkraidfs, mkhostfs

 13.49.  MKRAIDFS - script to create a raid filesystem


 �  SYNOPSIS: mkraidfs -r raidtype -n raidname -b backends [-c chunk]
    [-i iomode] [-q qlen] [-v] [-C capacity] [-S]

 �  DESCRIPTION: mkraidfs is a husky script which is used to perform
    all the necessary commands to create a Raid filesystem.

 �  OPTIONS:

 �  -r raidtype: Specify the raid type as raidtype for the raid set.
    Must be 0, 1, 3 or 5.

 �  -n raidname: Specify the name of the raid set as raidname.

 �  -b backends: Specify the comma separated list of the raid set's
    backends in the format used by rconf.

 �  -c iosize: Optionally specify the IOSIZE (in bytes) of the raid
    set.

 �  -i iomode: Optionally specify the raid set's iomode - read-write,
    read-only, write-only.

 �  -q qlen: Optionally specify the raid set's queue length for each
    backend.

 �  -v: Enable verbose mode which prints out the main actions (binding,
    engage commands) as they are performed.

 �  -C capacity: Optionally specify the raid set's size in 512-byte
    blocks.

 �  -S: Optionally specify that spares pool access is required should a
    backend fail.


    After parsing it's arguments mkraidfs creates the Raid Set's
    backend filesystems, typically, disks  (see  mkdisfs)  taking care
    of failed backends.  It then binds in the raid filesystem and
    engages the backends into the filesystem.  If spares access is
    requested, it enables the autorepair feature of the raid set.


 SEE ALSO: rconf, mkraidfs, mkhostfs, mkdiskfs, raid[0135]fs

 13.50.  MKSMON - script to start the scsi monitor daemon smon


 �  SYNOPSIS: mksmon controllerno hostport scsi_lun protocol_list

 �  DESCRIPTION: mksmon  is a husky script which is used to perform all
    the necessary commands to start the scsi monitor daemon smon given
    the controller number, hostport, scsi lun, and the block protocol
    list. Typically, mksmon, is run with it's arguments from the output
    of a mconf -list command.

 �  OPTIONS:

 �  controllerno: Specify the controller on which the scsi monitor
    daemon is to be run.

 �  hostport: Specify the host port through which the scsi monitor
    daemon communicates.

 �  scsi_lun: Specify the SCSI LUN the scsi monitor daemon is to
    respond to.

 �  protocol_list: Specify the comma separated block protocol list the
    scsi monitor daemon is to implement.


    After parsing it's arguments mksmon checks to see if it's already
    running and  issues  a  message  if  so  and exits. Otherwise, it
    creates the host filesystem (mkhostfs), creates a memory file and
    set of fifo's for smon to use and finally starts smon set.