Blame view

Documentation/blockdev/README.DAC960 33.9 KB
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
     Linux Driver for Mylex DAC960/AcceleRAID/eXtremeRAID PCI RAID Controllers
  
  			Version 2.2.11 for Linux 2.2.19
  			Version 2.4.11 for Linux 2.4.12
  
  			      PRODUCTION RELEASE
  
  				11 October 2001
  
  			       Leonard N. Zubkoff
  			       Dandelion Digital
  			       lnz@dandelion.com
  
  	 Copyright 1998-2001 by Leonard N. Zubkoff <lnz@dandelion.com>
  
  
  				 INTRODUCTION
  
  Mylex, Inc. designs and manufactures a variety of high performance PCI RAID
  controllers.  Mylex Corporation is located at 34551 Ardenwood Blvd., Fremont,
  California 94555, USA and can be reached at 510.796.6100 or on the World Wide
  Web at http://www.mylex.com.  Mylex Technical Support can be reached by
  electronic mail at mylexsup@us.ibm.com, by voice at 510.608.2400, or by FAX at
  510.745.7715.  Contact information for offices in Europe and Japan is available
  on their Web site.
  
  The latest information on Linux support for DAC960 PCI RAID Controllers, as
  well as the most recent release of this driver, will always be available from
  my Linux Home Page at URL "http://www.dandelion.com/Linux/".  The Linux DAC960
  driver supports all current Mylex PCI RAID controllers including the new
  eXtremeRAID 2000/3000 and AcceleRAID 352/170/160 models which have an entirely
  new firmware interface from the older eXtremeRAID 1100, AcceleRAID 150/200/250,
  and DAC960PJ/PG/PU/PD/PL.  See below for a complete controller list as well as
  minimum firmware version requirements.  For simplicity, in most places this
  documentation refers to DAC960 generically rather than explicitly listing all
  the supported models.
  
  Driver bug reports should be sent via electronic mail to "lnz@dandelion.com".
  Please include with the bug report the complete configuration messages reported
  by the driver at startup, along with any subsequent system messages relevant to
  the controller's operation, and a detailed description of your system's
  hardware configuration.  Driver bugs are actually quite rare; if you encounter
  problems with disks being marked offline, for example, please contact Mylex
  Technical Support as the problem is related to the hardware configuration
  rather than the Linux driver.
  
  Please consult the RAID controller documentation for detailed information
  regarding installation and configuration of the controllers.  This document
  primarily provides information specific to the Linux support.
  
  
  				DRIVER FEATURES
  
  The DAC960 RAID controllers are supported solely as high performance RAID
  controllers, not as interfaces to arbitrary SCSI devices.  The Linux DAC960
  driver operates at the block device level, the same level as the SCSI and IDE
  drivers.  Unlike other RAID controllers currently supported on Linux, the
  DAC960 driver is not dependent on the SCSI subsystem, and hence avoids all the
  complexity and unnecessary code that would be associated with an implementation
  as a SCSI driver.  The DAC960 driver is designed for as high a performance as
  possible with no compromises or extra code for compatibility with lower
  performance devices.  The DAC960 driver includes extensive error logging and
  online configuration management capabilities.  Except for initial configuration
  of the controller and adding new disk drives, most everything can be handled
  from Linux while the system is operational.
  
  The DAC960 driver is architected to support up to 8 controllers per system.
  Each DAC960 parallel SCSI controller can support up to 15 disk drives per
  channel, for a maximum of 60 drives on a four channel controller; the fibre
  channel eXtremeRAID 3000 controller supports up to 125 disk drives per loop for
  a total of 250 drives.  The drives installed on a controller are divided into
  one or more "Drive Groups", and then each Drive Group is subdivided further
  into 1 to 32 "Logical Drives".  Each Logical Drive has a specific RAID Level
  and caching policy associated with it, and it appears to Linux as a single
  block device.  Logical Drives are further subdivided into up to 7 partitions
  through the normal Linux and PC disk partitioning schemes.  Logical Drives are
  also known as "System Drives", and Drive Groups are also called "Packs".  Both
  terms are in use in the Mylex documentation; I have chosen to standardize on
  the more generic "Logical Drive" and "Drive Group".
890fbae28   Greg Kroah-Hartman   [PATCH] devfs: La...
80
81
82
  DAC960 RAID disk devices are named in the style of the obsolete Device File
  System (DEVFS).  The device corresponding to Logical Drive D on Controller C
  is referred to as /dev/rd/cCdD, and the partitions are called /dev/rd/cCdDp1
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
  through /dev/rd/cCdDp7.  For example, partition 3 of Logical Drive 5 on
  Controller 2 is referred to as /dev/rd/c2d5p3.  Note that unlike with SCSI
  disks the device names will not change in the event of a disk drive failure.
  The DAC960 driver is assigned major numbers 48 - 55 with one major number per
  controller.  The 8 bits of minor number are divided into 5 bits for the Logical
  Drive and 3 bits for the partition.
  
  
  	  SUPPORTED DAC960/AcceleRAID/eXtremeRAID PCI RAID CONTROLLERS
  
  The following list comprises the supported DAC960, AcceleRAID, and eXtremeRAID
  PCI RAID Controllers as of the date of this document.  It is recommended that
  anyone purchasing a Mylex PCI RAID Controller not in the following table
  contact the author beforehand to verify that it is or will be supported.
  
  eXtremeRAID 3000
  	    1 Wide Ultra-2/LVD SCSI channel
  	    2 External Fibre FC-AL channels
  	    233MHz StrongARM SA 110 Processor
  	    64 Bit 33MHz PCI (backward compatible with 32 Bit PCI slots)
  	    32MB/64MB ECC SDRAM Memory
  
  eXtremeRAID 2000
  	    4 Wide Ultra-160 LVD SCSI channels
  	    233MHz StrongARM SA 110 Processor
  	    64 Bit 33MHz PCI (backward compatible with 32 Bit PCI slots)
  	    32MB/64MB ECC SDRAM Memory
  
  AcceleRAID 352
  	    2 Wide Ultra-160 LVD SCSI channels
  	    100MHz Intel i960RN RISC Processor
  	    64 Bit 33MHz PCI (backward compatible with 32 Bit PCI slots)
  	    32MB/64MB ECC SDRAM Memory
  
  AcceleRAID 170
  	    1 Wide Ultra-160 LVD SCSI channel
  	    100MHz Intel i960RM RISC Processor
  	    16MB/32MB/64MB ECC SDRAM Memory
  
  AcceleRAID 160 (AcceleRAID 170LP)
  	    1 Wide Ultra-160 LVD SCSI channel
  	    100MHz Intel i960RS RISC Processor
  	    Built in 16M ECC SDRAM Memory
  	    PCI Low Profile Form Factor - fit for 2U height
  
  eXtremeRAID 1100 (DAC1164P)
  	    3 Wide Ultra-2/LVD SCSI channels
  	    233MHz StrongARM SA 110 Processor
  	    64 Bit 33MHz PCI (backward compatible with 32 Bit PCI slots)
  	    16MB/32MB/64MB Parity SDRAM Memory with Battery Backup
  
  AcceleRAID 250 (DAC960PTL1)
  	    Uses onboard Symbios SCSI chips on certain motherboards
  	    Also includes one onboard Wide Ultra-2/LVD SCSI Channel
  	    66MHz Intel i960RD RISC Processor
  	    4MB/8MB/16MB/32MB/64MB/128MB ECC EDO Memory
  
  AcceleRAID 200 (DAC960PTL0)
  	    Uses onboard Symbios SCSI chips on certain motherboards
  	    Includes no onboard SCSI Channels
  	    66MHz Intel i960RD RISC Processor
  	    4MB/8MB/16MB/32MB/64MB/128MB ECC EDO Memory
  
  AcceleRAID 150 (DAC960PRL)
  	    Uses onboard Symbios SCSI chips on certain motherboards
  	    Also includes one onboard Wide Ultra-2/LVD SCSI Channel
  	    33MHz Intel i960RP RISC Processor
  	    4MB Parity EDO Memory
  
  DAC960PJ    1/2/3 Wide Ultra SCSI-3 Channels
  	    66MHz Intel i960RD RISC Processor
  	    4MB/8MB/16MB/32MB/64MB/128MB ECC EDO Memory
  
  DAC960PG    1/2/3 Wide Ultra SCSI-3 Channels
  	    33MHz Intel i960RP RISC Processor
  	    4MB/8MB ECC EDO Memory
  
  DAC960PU    1/2/3 Wide Ultra SCSI-3 Channels
  	    Intel i960CF RISC Processor
  	    4MB/8MB EDRAM or 2MB/4MB/8MB/16MB/32MB DRAM Memory
  
  DAC960PD    1/2/3 Wide Fast SCSI-2 Channels
  	    Intel i960CF RISC Processor
  	    4MB/8MB EDRAM or 2MB/4MB/8MB/16MB/32MB DRAM Memory
  
  DAC960PL    1/2/3 Wide Fast SCSI-2 Channels
  	    Intel i960 RISC Processor
  	    2MB/4MB/8MB/16MB/32MB DRAM Memory
  
  DAC960P	    1/2/3 Wide Fast SCSI-2 Channels
  	    Intel i960 RISC Processor
  	    2MB/4MB/8MB/16MB/32MB DRAM Memory
  
  For the eXtremeRAID 2000/3000 and AcceleRAID 352/170/160, firmware version
  6.00-01 or above is required.
  
  For the eXtremeRAID 1100, firmware version 5.06-0-52 or above is required.
  
  For the AcceleRAID 250, 200, and 150, firmware version 4.06-0-57 or above is
  required.
  
  For the DAC960PJ and DAC960PG, firmware version 4.06-0-00 or above is required.
  
  For the DAC960PU, DAC960PD, DAC960PL, and DAC960P, either firmware version
  3.51-0-04 or above is required (for dual Flash ROM controllers), or firmware
  version 2.73-0-00 or above is required (for single Flash ROM controllers)
  
  Please note that not all SCSI disk drives are suitable for use with DAC960
  controllers, and only particular firmware versions of any given model may
  actually function correctly.  Similarly, not all motherboards have a BIOS that
  properly initializes the AcceleRAID 250, AcceleRAID 200, AcceleRAID 150,
  DAC960PJ, and DAC960PG because the Intel i960RD/RP is a multi-function device.
  If in doubt, contact Mylex RAID Technical Support (mylexsup@us.ibm.com) to
  verify compatibility.  Mylex makes available a hard disk compatibility list at
  http://www.mylex.com/support/hdcomp/hd-lists.html.
  
  
  			      DRIVER INSTALLATION
  
  This distribution was prepared for Linux kernel version 2.2.19 or 2.4.12.
  
  To install the DAC960 RAID driver, you may use the following commands,
  replacing "/usr/src" with wherever you keep your Linux kernel source tree:
  
    cd /usr/src
    tar -xvzf DAC960-2.2.11.tar.gz (or DAC960-2.4.11.tar.gz)
    mv README.DAC960 linux/Documentation
    mv DAC960.[ch] linux/drivers/block
    patch -p0 < DAC960.patch (if DAC960.patch is included)
    cd linux
    make config
    make bzImage (or zImage)
25eb650a6   Wanlong Gao   doc: fix wrong ar...
215
  Then install "arch/x86/boot/bzImage" or "arch/x86/boot/zImage" as your
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
  standard kernel, run lilo if appropriate, and reboot.
  
  To create the necessary devices in /dev, the "make_rd" script included in
  "DAC960-Utilities.tar.gz" from http://www.dandelion.com/Linux/ may be used.
  LILO 21 and FDISK v2.9 include DAC960 support; also included in this archive
  are patches to LILO 20 and FDISK v2.8 that add DAC960 support, along with
  statically linked executables of LILO and FDISK.  This modified version of LILO
  will allow booting from a DAC960 controller and/or mounting the root file
  system from a DAC960.
  
  Red Hat Linux 6.0 and SuSE Linux 6.1 include support for Mylex PCI RAID
  controllers.  Installing directly onto a DAC960 may be problematic from other
  Linux distributions until their installation utilities are updated.
  
  
  			      INSTALLATION NOTES
  
  Before installing Linux or adding DAC960 logical drives to an existing Linux
  system, the controller must first be configured to provide one or more logical
  drives using the BIOS Configuration Utility or DACCF.  Please note that since
  there are only at most 6 usable partitions on each logical drive, systems
  requiring more partitions should subdivide a drive group into multiple logical
  drives, each of which can have up to 6 usable partitions.  Also, note that with
  large disk arrays it is advisable to enable the 8GB BIOS Geometry (255/63)
  rather than accepting the default 2GB BIOS Geometry (128/32); failing to so do
  will cause the logical drive geometry to have more than 65535 cylinders which
  will make it impossible for FDISK to be used properly.  The 8GB BIOS Geometry
  can be enabled by configuring the DAC960 BIOS, which is accessible via Alt-M
  during the BIOS initialization sequence.
  
  For maximum performance and the most efficient E2FSCK performance, it is
  recommended that EXT2 file systems be built with a 4KB block size and 16 block
  stride to match the DAC960 controller's 64KB default stripe size.  The command
  "mke2fs -b 4096 -R stride=16 <device>" is appropriate.  Unless there will be a
  large number of small files on the file systems, it is also beneficial to add
  the "-i 16384" option to increase the bytes per inode parameter thereby
  reducing the file system metadata.  Finally, on systems that will only be run
  with Linux 2.2 or later kernels it is beneficial to enable sparse superblocks
  with the "-s 1" option.
  
  
  		      DAC960 ANNOUNCEMENTS MAILING LIST
  
  The DAC960 Announcements Mailing List provides a forum for informing Linux
  users of new driver releases and other announcements regarding Linux support
  for DAC960 PCI RAID Controllers.  To join the mailing list, send a message to
  "dac960-announce-request@dandelion.com" with the line "subscribe" in the
  message body.
  
  
  		CONTROLLER CONFIGURATION AND STATUS MONITORING
  
  The DAC960 RAID controllers running firmware 4.06 or above include a Background
  Initialization facility so that system downtime is minimized both for initial
  installation and subsequent configuration of additional storage.  The BIOS
  Configuration Utility (accessible via Alt-R during the BIOS initialization
  sequence) is used to quickly configure the controller, and then the logical
  drives that have been created are available for immediate use even while they
  are still being initialized by the controller.  The primary need for online
  configuration and status monitoring is then to avoid system downtime when disk
  drives fail and must be replaced.  Mylex's online monitoring and configuration
  utilities are being ported to Linux and will become available at some point in
  the future.  Note that with a SAF-TE (SCSI Accessed Fault-Tolerant Enclosure)
  enclosure, the controller is able to rebuild failed drives automatically as
  soon as a drive replacement is made available.
  
  The primary interfaces for controller configuration and status monitoring are
  special files created in the /proc/rd/... hierarchy along with the normal
  system console logging mechanism.  Whenever the system is operating, the DAC960
  driver queries each controller for status information every 10 seconds, and
  checks for additional conditions every 60 seconds.  The initial status of each
  controller is always available for controller N in /proc/rd/cN/initial_status,
  and the current status as of the last status monitoring query is available in
  /proc/rd/cN/current_status.  In addition, status changes are also logged by the
  driver to the system console and will appear in the log files maintained by
  syslog.  The progress of asynchronous rebuild or consistency check operations
  is also available in /proc/rd/cN/current_status, and progress messages are
  logged to the system console at most every 60 seconds.
  
  Starting with the 2.2.3/2.0.3 versions of the driver, the status information
  available in /proc/rd/cN/initial_status and /proc/rd/cN/current_status has been
  augmented to include the vendor, model, revision, and serial number (if
  available) for each physical device found connected to the controller:
  
  ***** DAC960 RAID Driver Version 2.2.3 of 19 August 1999 *****
  Copyright 1998-1999 by Leonard N. Zubkoff <lnz@dandelion.com>
  Configuring Mylex DAC960PRL PCI RAID Controller
    Firmware Version: 4.07-0-07, Channels: 1, Memory Size: 16MB
    PCI Bus: 1, Device: 4, Function: 1, I/O Address: Unassigned
    PCI Address: 0xFE300000 mapped at 0xA0800000, IRQ Channel: 21
    Controller Queue Depth: 128, Maximum Blocks per Command: 128
    Driver Queue Depth: 127, Maximum Scatter/Gather Segments: 33
    Stripe Size: 64KB, Segment Size: 8KB, BIOS Geometry: 255/63
    SAF-TE Enclosure Management Enabled
    Physical Devices:
      0:0  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       68016775HA
           Disk Status: Online, 17928192 blocks
      0:1  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       68004E53HA
           Disk Status: Online, 17928192 blocks
      0:2  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       13013935HA
           Disk Status: Online, 17928192 blocks
      0:3  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       13016897HA
           Disk Status: Online, 17928192 blocks
      0:4  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       68019905HA
           Disk Status: Online, 17928192 blocks
      0:5  Vendor: IBM       Model: DRVS09D           Revision: 0270
           Serial Number:       68012753HA
           Disk Status: Online, 17928192 blocks
      0:6  Vendor: ESG-SHV   Model: SCA HSBP M6       Revision: 0.61
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 89640960 blocks, Write Thru
    No Rebuild or Consistency Check in Progress
  
  To simplify the monitoring process for custom software, the special file
  /proc/rd/status returns "OK" when all DAC960 controllers in the system are
  operating normally and no failures have occurred, or "ALERT" if any logical
  drives are offline or critical or any non-standby physical drives are dead.
  
  Configuration commands for controller N are available via the special file
  /proc/rd/cN/user_command.  A human readable command can be written to this
  special file to initiate a configuration operation, and the results of the
  operation can then be read back from the special file in addition to being
  logged to the system console.  The shell command sequence
  
    echo "<configuration-command>" > /proc/rd/c0/user_command
    cat /proc/rd/c0/user_command
  
  is typically used to execute configuration commands.  The configuration
  commands are:
  
    flush-cache
  
      The "flush-cache" command flushes the controller's cache.  The system
      automatically flushes the cache at shutdown or if the driver module is
      unloaded, so this command is only needed to be certain a write back cache
      is flushed to disk before the system is powered off by a command to a UPS.
      Note that the flush-cache command also stops an asynchronous rebuild or
      consistency check, so it should not be used except when the system is being
      halted.
  
    kill <channel>:<target-id>
  
      The "kill" command marks the physical drive <channel>:<target-id> as DEAD.
      This command is provided primarily for testing, and should not be used
      during normal system operation.
  
    make-online <channel>:<target-id>
  
      The "make-online" command changes the physical drive <channel>:<target-id>
      from status DEAD to status ONLINE.  In cases where multiple physical drives
      have been killed simultaneously, this command may be used to bring all but
      one of them back online, after which a rebuild to the final drive is
      necessary.
  
      Warning: make-online should only be used on a dead physical drive that is
      an active part of a drive group, never on a standby drive.  The command
      should never be used on a dead drive that is part of a critical logical
      drive; rebuild should be used if only a single drive is dead.
  
    make-standby <channel>:<target-id>
  
      The "make-standby" command changes physical drive <channel>:<target-id>
      from status DEAD to status STANDBY.  It should only be used in cases where
      a dead drive was replaced after an automatic rebuild was performed onto a
      standby drive.  It cannot be used to add a standby drive to the controller
      configuration if one was not created initially; the BIOS Configuration
      Utility must be used for that currently.
  
    rebuild <channel>:<target-id>
  
      The "rebuild" command initiates an asynchronous rebuild onto physical drive
      <channel>:<target-id>.  It should only be used when a dead drive has been
      replaced.
  
    check-consistency <logical-drive-number>
  
      The "check-consistency" command initiates an asynchronous consistency check
      of <logical-drive-number> with automatic restoration.  It can be used
      whenever it is desired to verify the consistency of the redundancy
      information.
  
    cancel-rebuild
    cancel-consistency-check
  
      The "cancel-rebuild" and "cancel-consistency-check" commands cancel any
      rebuild or consistency check operations previously initiated.
  
  
  	       EXAMPLE I - DRIVE FAILURE WITHOUT A STANDBY DRIVE
  
  The following annotated logs demonstrate the controller configuration and and
  online status monitoring capabilities of the Linux DAC960 Driver.  The test
  configuration comprises 6 1GB Quantum Atlas I disk drives on two channels of a
  DAC960PJ controller.  The physical drives are configured into a single drive
  group without a standby drive, and the drive group has been configured into two
  logical drives, one RAID-5 and one RAID-6.  Note that these logs are from an
  earlier version of the driver and the messages have changed somewhat with newer
  releases, but the functionality remains similar.  First, here is the current
  status of the RAID configuration:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
  ***** DAC960 RAID Driver Version 2.0.0 of 23 March 1999 *****
  Copyright 1998-1999 by Leonard N. Zubkoff <lnz@dandelion.com>
  Configuring Mylex DAC960PJ PCI RAID Controller
    Firmware Version: 4.06-0-08, Channels: 3, Memory Size: 8MB
    PCI Bus: 0, Device: 19, Function: 1, I/O Address: Unassigned
    PCI Address: 0xFD4FC000 mapped at 0x8807000, IRQ Channel: 9
    Controller Queue Depth: 128, Maximum Blocks per Command: 128
    Driver Queue Depth: 127, Maximum Scatter/Gather Segments: 33
    Stripe Size: 64KB, Segment Size: 8KB, BIOS Geometry: 255/63
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 5498880 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Online, 3305472 blocks, Write Thru
    No Rebuild or Consistency Check in Progress
  
  gwynedd:/u/lnz# cat /proc/rd/status
  OK
  
  The above messages indicate that everything is healthy, and /proc/rd/status
  returns "OK" indicating that there are no problems with any DAC960 controller
  in the system.  For demonstration purposes, while I/O is active Physical Drive
  1:1 is now disconnected, simulating a drive failure.  The failure is noted by
  the driver within 10 seconds of the controller's having detected it, and the
  driver logs the following console status messages indicating that Logical
  Drives 0 and 1 are now CRITICAL as a result of Physical Drive 1:1 being DEAD:
  
  DAC960#0: Physical Drive 1:2 Error Log: Sense Key = 6, ASC = 29, ASCQ = 02
  DAC960#0: Physical Drive 1:3 Error Log: Sense Key = 6, ASC = 29, ASCQ = 02
  DAC960#0: Physical Drive 1:1 killed because of timeout on SCSI command
  DAC960#0: Physical Drive 1:1 is now DEAD
  DAC960#0: Logical Drive 0 (/dev/rd/c0d0) is now CRITICAL
  DAC960#0: Logical Drive 1 (/dev/rd/c0d1) is now CRITICAL
  
  The Sense Keys logged here are just Check Condition / Unit Attention conditions
  arising from a SCSI bus reset that is forced by the controller during its error
  recovery procedures.  Concurrently with the above, the driver status available
  from /proc/rd also reflects the drive failure.  The status message in
  /proc/rd/status has changed from "OK" to "ALERT":
  
  gwynedd:/u/lnz# cat /proc/rd/status
  ALERT
  
  and /proc/rd/c0/current_status has been updated:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Dead, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Critical, 5498880 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Critical, 3305472 blocks, Write Thru
    No Rebuild or Consistency Check in Progress
  
  Since there are no standby drives configured, the system can continue to access
  the logical drives in a performance degraded mode until the failed drive is
  replaced and a rebuild operation completed to restore the redundancy of the
  logical drives.  Once Physical Drive 1:1 is replaced with a properly
  functioning drive, or if the physical drive was killed without having failed
  (e.g., due to electrical problems on the SCSI bus), the user can instruct the
  controller to initiate a rebuild operation onto the newly replaced drive:
  
  gwynedd:/u/lnz# echo "rebuild 1:1" > /proc/rd/c0/user_command
  gwynedd:/u/lnz# cat /proc/rd/c0/user_command
  Rebuild of Physical Drive 1:1 Initiated
  
  The echo command instructs the controller to initiate an asynchronous rebuild
  operation onto Physical Drive 1:1, and the status message that results from the
  operation is then available for reading from /proc/rd/c0/user_command, as well
  as being logged to the console by the driver.
  
  Within 10 seconds of this command the driver logs the initiation of the
  asynchronous rebuild operation:
  
  DAC960#0: Rebuild of Physical Drive 1:1 Initiated
  DAC960#0: Physical Drive 1:1 Error Log: Sense Key = 6, ASC = 29, ASCQ = 01
  DAC960#0: Physical Drive 1:1 is now WRITE-ONLY
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 1% completed
  
  and /proc/rd/c0/current_status is updated:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Write-Only, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Critical, 5498880 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Critical, 3305472 blocks, Write Thru
    Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 6% completed
  
  As the rebuild progresses, the current status in /proc/rd/c0/current_status is
  updated every 10 seconds:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Write-Only, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Critical, 5498880 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Critical, 3305472 blocks, Write Thru
    Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 15% completed
  
  and every minute a progress message is logged to the console by the driver:
  
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 32% completed
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 63% completed
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 94% completed
  DAC960#0: Rebuild in Progress: Logical Drive 1 (/dev/rd/c0d1) 94% completed
  
  Finally, the rebuild completes successfully.  The driver logs the status of the 
  logical and physical drives and the rebuild completion:
  
  DAC960#0: Rebuild Completed Successfully
  DAC960#0: Physical Drive 1:1 is now ONLINE
  DAC960#0: Logical Drive 0 (/dev/rd/c0d0) is now ONLINE
  DAC960#0: Logical Drive 1 (/dev/rd/c0d1) is now ONLINE
  
  /proc/rd/c0/current_status is updated:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 5498880 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Online, 3305472 blocks, Write Thru
    Rebuild Completed Successfully
  
  and /proc/rd/status indicates that everything is healthy once again:
  
  gwynedd:/u/lnz# cat /proc/rd/status
  OK
  
  
  		EXAMPLE II - DRIVE FAILURE WITH A STANDBY DRIVE
  
  The following annotated logs demonstrate the controller configuration and and
  online status monitoring capabilities of the Linux DAC960 Driver.  The test
  configuration comprises 6 1GB Quantum Atlas I disk drives on two channels of a
  DAC960PJ controller.  The physical drives are configured into a single drive
  group with a standby drive, and the drive group has been configured into two
  logical drives, one RAID-5 and one RAID-6.  Note that these logs are from an
  earlier version of the driver and the messages have changed somewhat with newer
  releases, but the functionality remains similar.  First, here is the current
  status of the RAID configuration:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
  ***** DAC960 RAID Driver Version 2.0.0 of 23 March 1999 *****
  Copyright 1998-1999 by Leonard N. Zubkoff <lnz@dandelion.com>
  Configuring Mylex DAC960PJ PCI RAID Controller
    Firmware Version: 4.06-0-08, Channels: 3, Memory Size: 8MB
    PCI Bus: 0, Device: 19, Function: 1, I/O Address: Unassigned
    PCI Address: 0xFD4FC000 mapped at 0x8807000, IRQ Channel: 9
    Controller Queue Depth: 128, Maximum Blocks per Command: 128
    Driver Queue Depth: 127, Maximum Scatter/Gather Segments: 33
    Stripe Size: 64KB, Segment Size: 8KB, BIOS Geometry: 255/63
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Online, 2201600 blocks
      1:3 - Disk: Standby, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 4399104 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Online, 2754560 blocks, Write Thru
    No Rebuild or Consistency Check in Progress
  
  gwynedd:/u/lnz# cat /proc/rd/status
  OK
  
  The above messages indicate that everything is healthy, and /proc/rd/status
  returns "OK" indicating that there are no problems with any DAC960 controller
  in the system.  For demonstration purposes, while I/O is active Physical Drive
  1:2 is now disconnected, simulating a drive failure.  The failure is noted by
  the driver within 10 seconds of the controller's having detected it, and the
  driver logs the following console status messages:
  
  DAC960#0: Physical Drive 1:1 Error Log: Sense Key = 6, ASC = 29, ASCQ = 02
  DAC960#0: Physical Drive 1:3 Error Log: Sense Key = 6, ASC = 29, ASCQ = 02
  DAC960#0: Physical Drive 1:2 killed because of timeout on SCSI command
  DAC960#0: Physical Drive 1:2 is now DEAD
  DAC960#0: Physical Drive 1:2 killed because it was removed
  DAC960#0: Logical Drive 0 (/dev/rd/c0d0) is now CRITICAL
  DAC960#0: Logical Drive 1 (/dev/rd/c0d1) is now CRITICAL
  
  Since a standby drive is configured, the controller automatically begins
  rebuilding onto the standby drive:
  
  DAC960#0: Physical Drive 1:3 is now WRITE-ONLY
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 4% completed
  
  Concurrently with the above, the driver status available from /proc/rd also
  reflects the drive failure and automatic rebuild.  The status message in
  /proc/rd/status has changed from "OK" to "ALERT":
  
  gwynedd:/u/lnz# cat /proc/rd/status
  ALERT
  
  and /proc/rd/c0/current_status has been updated:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Dead, 2201600 blocks
      1:3 - Disk: Write-Only, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Critical, 4399104 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Critical, 2754560 blocks, Write Thru
    Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 4% completed
  
  As the rebuild progresses, the current status in /proc/rd/c0/current_status is
  updated every 10 seconds:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Dead, 2201600 blocks
      1:3 - Disk: Write-Only, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Critical, 4399104 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Critical, 2754560 blocks, Write Thru
    Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 40% completed
  
  and every minute a progress message is logged on the console by the driver:
  
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 40% completed
  DAC960#0: Rebuild in Progress: Logical Drive 0 (/dev/rd/c0d0) 76% completed
  DAC960#0: Rebuild in Progress: Logical Drive 1 (/dev/rd/c0d1) 66% completed
  DAC960#0: Rebuild in Progress: Logical Drive 1 (/dev/rd/c0d1) 84% completed
  
  Finally, the rebuild completes successfully.  The driver logs the status of the 
  logical and physical drives and the rebuild completion:
  
  DAC960#0: Rebuild Completed Successfully
  DAC960#0: Physical Drive 1:3 is now ONLINE
  DAC960#0: Logical Drive 0 (/dev/rd/c0d0) is now ONLINE
  DAC960#0: Logical Drive 1 (/dev/rd/c0d1) is now ONLINE
  
  /proc/rd/c0/current_status is updated:
  
  ***** DAC960 RAID Driver Version 2.0.0 of 23 March 1999 *****
  Copyright 1998-1999 by Leonard N. Zubkoff <lnz@dandelion.com>
  Configuring Mylex DAC960PJ PCI RAID Controller
    Firmware Version: 4.06-0-08, Channels: 3, Memory Size: 8MB
    PCI Bus: 0, Device: 19, Function: 1, I/O Address: Unassigned
    PCI Address: 0xFD4FC000 mapped at 0x8807000, IRQ Channel: 9
    Controller Queue Depth: 128, Maximum Blocks per Command: 128
    Driver Queue Depth: 127, Maximum Scatter/Gather Segments: 33
    Stripe Size: 64KB, Segment Size: 8KB, BIOS Geometry: 255/63
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Dead, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 4399104 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Online, 2754560 blocks, Write Thru
    Rebuild Completed Successfully
  
  and /proc/rd/status indicates that everything is healthy once again:
  
  gwynedd:/u/lnz# cat /proc/rd/status
  OK
  
  Note that the absence of a viable standby drive does not create an "ALERT"
  status.  Once dead Physical Drive 1:2 has been replaced, the controller must be
  told that this has occurred and that the newly replaced drive should become the
  new standby drive:
  
  gwynedd:/u/lnz# echo "make-standby 1:2" > /proc/rd/c0/user_command
  gwynedd:/u/lnz# cat /proc/rd/c0/user_command
  Make Standby of Physical Drive 1:2 Succeeded
  
  The echo command instructs the controller to make Physical Drive 1:2 into a
  standby drive, and the status message that results from the operation is then
  available for reading from /proc/rd/c0/user_command, as well as being logged to
  the console by the driver.  Within 60 seconds of this command the driver logs:
  
  DAC960#0: Physical Drive 1:2 Error Log: Sense Key = 6, ASC = 29, ASCQ = 01
  DAC960#0: Physical Drive 1:2 is now STANDBY
  DAC960#0: Make Standby of Physical Drive 1:2 Succeeded
  
  and /proc/rd/c0/current_status is updated:
  
  gwynedd:/u/lnz# cat /proc/rd/c0/current_status
    ...
    Physical Devices:
      0:1 - Disk: Online, 2201600 blocks
      0:2 - Disk: Online, 2201600 blocks
      0:3 - Disk: Online, 2201600 blocks
      1:1 - Disk: Online, 2201600 blocks
      1:2 - Disk: Standby, 2201600 blocks
      1:3 - Disk: Online, 2201600 blocks
    Logical Drives:
      /dev/rd/c0d0: RAID-5, Online, 4399104 blocks, Write Thru
      /dev/rd/c0d1: RAID-6, Online, 2754560 blocks, Write Thru
    Rebuild Completed Successfully