cciss.c revision ef7822c2fbbb004e30e1fb6f79d6f8a1a1a144d7
1/*
2 *    Disk Array driver for HP Smart Array controllers.
3 *    (C) Copyright 2000, 2007 Hewlett-Packard Development Company, L.P.
4 *
5 *    This program is free software; you can redistribute it and/or modify
6 *    it under the terms of the GNU General Public License as published by
7 *    the Free Software Foundation; version 2 of the License.
8 *
9 *    This program is distributed in the hope that it will be useful,
10 *    but WITHOUT ANY WARRANTY; without even the implied warranty of
11 *    MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. See the GNU
12 *    General Public License for more details.
13 *
14 *    You should have received a copy of the GNU General Public License
15 *    along with this program; if not, write to the Free Software
16 *    Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
17 *    02111-1307, USA.
18 *
19 *    Questions/Comments/Bugfixes to iss_storagedev@hp.com
20 *
21 */
22
23#include <linux/module.h>
24#include <linux/interrupt.h>
25#include <linux/types.h>
26#include <linux/pci.h>
27#include <linux/kernel.h>
28#include <linux/slab.h>
29#include <linux/delay.h>
30#include <linux/major.h>
31#include <linux/fs.h>
32#include <linux/bio.h>
33#include <linux/blkpg.h>
34#include <linux/timer.h>
35#include <linux/proc_fs.h>
36#include <linux/seq_file.h>
37#include <linux/init.h>
38#include <linux/hdreg.h>
39#include <linux/spinlock.h>
40#include <linux/compat.h>
41#include <linux/blktrace_api.h>
42#include <asm/uaccess.h>
43#include <asm/io.h>
44
45#include <linux/dma-mapping.h>
46#include <linux/blkdev.h>
47#include <linux/genhd.h>
48#include <linux/completion.h>
49#include <scsi/scsi.h>
50#include <scsi/sg.h>
51#include <scsi/scsi_ioctl.h>
52#include <linux/cdrom.h>
53#include <linux/scatterlist.h>
54
55#define CCISS_DRIVER_VERSION(maj,min,submin) ((maj<<16)|(min<<8)|(submin))
56#define DRIVER_NAME "HP CISS Driver (v 3.6.20)"
57#define DRIVER_VERSION CCISS_DRIVER_VERSION(3, 6, 20)
58
59/* Embedded module documentation macros - see modules.h */
60MODULE_AUTHOR("Hewlett-Packard Company");
61MODULE_DESCRIPTION("Driver for HP Smart Array Controllers");
62MODULE_SUPPORTED_DEVICE("HP SA5i SA5i+ SA532 SA5300 SA5312 SA641 SA642 SA6400"
63			" SA6i P600 P800 P400 P400i E200 E200i E500 P700m"
64			" Smart Array G2 Series SAS/SATA Controllers");
65MODULE_VERSION("3.6.20");
66MODULE_LICENSE("GPL");
67
68#include "cciss_cmd.h"
69#include "cciss.h"
70#include <linux/cciss_ioctl.h>
71
72/* define the PCI info for the cards we can control */
73static const struct pci_device_id cciss_pci_device_id[] = {
74	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISS,  0x0E11, 0x4070},
75	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSB, 0x0E11, 0x4080},
76	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSB, 0x0E11, 0x4082},
77	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSB, 0x0E11, 0x4083},
78	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSC, 0x0E11, 0x4091},
79	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSC, 0x0E11, 0x409A},
80	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSC, 0x0E11, 0x409B},
81	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSC, 0x0E11, 0x409C},
82	{PCI_VENDOR_ID_COMPAQ, PCI_DEVICE_ID_COMPAQ_CISSC, 0x0E11, 0x409D},
83	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSA,     0x103C, 0x3225},
84	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSC,     0x103C, 0x3223},
85	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSC,     0x103C, 0x3234},
86	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSC,     0x103C, 0x3235},
87	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSD,     0x103C, 0x3211},
88	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSD,     0x103C, 0x3212},
89	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSD,     0x103C, 0x3213},
90	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSD,     0x103C, 0x3214},
91	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSD,     0x103C, 0x3215},
92	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSC,     0x103C, 0x3237},
93	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSC,     0x103C, 0x323D},
94	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSE,     0x103C, 0x3241},
95	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSE,     0x103C, 0x3243},
96	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSE,     0x103C, 0x3245},
97	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSE,     0x103C, 0x3247},
98	{PCI_VENDOR_ID_HP,     PCI_DEVICE_ID_HP_CISSE,     0x103C, 0x3249},
99	{PCI_VENDOR_ID_HP,     PCI_ANY_ID,	PCI_ANY_ID, PCI_ANY_ID,
100		PCI_CLASS_STORAGE_RAID << 8, 0xffff << 8, 0},
101	{0,}
102};
103
104MODULE_DEVICE_TABLE(pci, cciss_pci_device_id);
105
106/*  board_id = Subsystem Device ID & Vendor ID
107 *  product = Marketing Name for the board
108 *  access = Address of the struct of function pointers
109 */
110static struct board_type products[] = {
111	{0x40700E11, "Smart Array 5300", &SA5_access},
112	{0x40800E11, "Smart Array 5i", &SA5B_access},
113	{0x40820E11, "Smart Array 532", &SA5B_access},
114	{0x40830E11, "Smart Array 5312", &SA5B_access},
115	{0x409A0E11, "Smart Array 641", &SA5_access},
116	{0x409B0E11, "Smart Array 642", &SA5_access},
117	{0x409C0E11, "Smart Array 6400", &SA5_access},
118	{0x409D0E11, "Smart Array 6400 EM", &SA5_access},
119	{0x40910E11, "Smart Array 6i", &SA5_access},
120	{0x3225103C, "Smart Array P600", &SA5_access},
121	{0x3223103C, "Smart Array P800", &SA5_access},
122	{0x3234103C, "Smart Array P400", &SA5_access},
123	{0x3235103C, "Smart Array P400i", &SA5_access},
124	{0x3211103C, "Smart Array E200i", &SA5_access},
125	{0x3212103C, "Smart Array E200", &SA5_access},
126	{0x3213103C, "Smart Array E200i", &SA5_access},
127	{0x3214103C, "Smart Array E200i", &SA5_access},
128	{0x3215103C, "Smart Array E200i", &SA5_access},
129	{0x3237103C, "Smart Array E500", &SA5_access},
130	{0x323D103C, "Smart Array P700m", &SA5_access},
131	{0x3241103C, "Smart Array P212", &SA5_access},
132	{0x3243103C, "Smart Array P410", &SA5_access},
133	{0x3245103C, "Smart Array P410i", &SA5_access},
134	{0x3247103C, "Smart Array P411", &SA5_access},
135	{0x3249103C, "Smart Array P812", &SA5_access},
136	{0xFFFF103C, "Unknown Smart Array", &SA5_access},
137};
138
139/* How long to wait (in milliseconds) for board to go into simple mode */
140#define MAX_CONFIG_WAIT 30000
141#define MAX_IOCTL_CONFIG_WAIT 1000
142
143/*define how many times we will try a command because of bus resets */
144#define MAX_CMD_RETRIES 3
145
146#define MAX_CTLR	32
147
148/* Originally cciss driver only supports 8 major numbers */
149#define MAX_CTLR_ORIG 	8
150
151static ctlr_info_t *hba[MAX_CTLR];
152
153static void do_cciss_request(struct request_queue *q);
154static irqreturn_t do_cciss_intr(int irq, void *dev_id);
155static int cciss_open(struct block_device *bdev, fmode_t mode);
156static int cciss_release(struct gendisk *disk, fmode_t mode);
157static int cciss_ioctl(struct block_device *bdev, fmode_t mode,
158		       unsigned int cmd, unsigned long arg);
159static int cciss_getgeo(struct block_device *bdev, struct hd_geometry *geo);
160
161static int cciss_revalidate(struct gendisk *disk);
162static int rebuild_lun_table(ctlr_info_t *h, int first_time);
163static int deregister_disk(struct gendisk *disk, drive_info_struct *drv,
164			   int clear_all);
165
166static void cciss_read_capacity(int ctlr, int logvol, int withirq,
167			sector_t *total_size, unsigned int *block_size);
168static void cciss_read_capacity_16(int ctlr, int logvol, int withirq,
169			sector_t *total_size, unsigned int *block_size);
170static void cciss_geometry_inquiry(int ctlr, int logvol,
171			int withirq, sector_t total_size,
172			unsigned int block_size, InquiryData_struct *inq_buff,
173				   drive_info_struct *drv);
174static void __devinit cciss_interrupt_mode(ctlr_info_t *, struct pci_dev *,
175					   __u32);
176static void start_io(ctlr_info_t *h);
177static int sendcmd(__u8 cmd, int ctlr, void *buff, size_t size,
178		   unsigned int use_unit_num, unsigned int log_unit,
179		   __u8 page_code, unsigned char *scsi3addr, int cmd_type);
180static int sendcmd_withirq(__u8 cmd, int ctlr, void *buff, size_t size,
181			   unsigned int use_unit_num, unsigned int log_unit,
182			   __u8 page_code, int cmd_type);
183
184static void fail_all_cmds(unsigned long ctlr);
185
186#ifdef CONFIG_PROC_FS
187static void cciss_procinit(int i);
188#else
189static void cciss_procinit(int i)
190{
191}
192#endif				/* CONFIG_PROC_FS */
193
194#ifdef CONFIG_COMPAT
195static int cciss_compat_ioctl(struct block_device *, fmode_t,
196			      unsigned, unsigned long);
197#endif
198
199static struct block_device_operations cciss_fops = {
200	.owner = THIS_MODULE,
201	.open = cciss_open,
202	.release = cciss_release,
203	.locked_ioctl = cciss_ioctl,
204	.getgeo = cciss_getgeo,
205#ifdef CONFIG_COMPAT
206	.compat_ioctl = cciss_compat_ioctl,
207#endif
208	.revalidate_disk = cciss_revalidate,
209};
210
211/*
212 * Enqueuing and dequeuing functions for cmdlists.
213 */
214static inline void addQ(CommandList_struct **Qptr, CommandList_struct *c)
215{
216	if (*Qptr == NULL) {
217		*Qptr = c;
218		c->next = c->prev = c;
219	} else {
220		c->prev = (*Qptr)->prev;
221		c->next = (*Qptr);
222		(*Qptr)->prev->next = c;
223		(*Qptr)->prev = c;
224	}
225}
226
227static inline CommandList_struct *removeQ(CommandList_struct **Qptr,
228					  CommandList_struct *c)
229{
230	if (c && c->next != c) {
231		if (*Qptr == c)
232			*Qptr = c->next;
233		c->prev->next = c->next;
234		c->next->prev = c->prev;
235	} else {
236		*Qptr = NULL;
237	}
238	return c;
239}
240
241#include "cciss_scsi.c"		/* For SCSI tape support */
242
243#define RAID_UNKNOWN 6
244
245#ifdef CONFIG_PROC_FS
246
247/*
248 * Report information about this controller.
249 */
250#define ENG_GIG 1000000000
251#define ENG_GIG_FACTOR (ENG_GIG/512)
252#define ENGAGE_SCSI	"engage scsi"
253static const char *raid_label[] = { "0", "4", "1(1+0)", "5", "5+1", "ADG",
254	"UNKNOWN"
255};
256
257static struct proc_dir_entry *proc_cciss;
258
259static void cciss_seq_show_header(struct seq_file *seq)
260{
261	ctlr_info_t *h = seq->private;
262
263	seq_printf(seq, "%s: HP %s Controller\n"
264		"Board ID: 0x%08lx\n"
265		"Firmware Version: %c%c%c%c\n"
266		"IRQ: %d\n"
267		"Logical drives: %d\n"
268		"Current Q depth: %d\n"
269		"Current # commands on controller: %d\n"
270		"Max Q depth since init: %d\n"
271		"Max # commands on controller since init: %d\n"
272		"Max SG entries since init: %d\n",
273		h->devname,
274		h->product_name,
275		(unsigned long)h->board_id,
276		h->firm_ver[0], h->firm_ver[1], h->firm_ver[2],
277		h->firm_ver[3], (unsigned int)h->intr[SIMPLE_MODE_INT],
278		h->num_luns,
279		h->Qdepth, h->commands_outstanding,
280		h->maxQsinceinit, h->max_outstanding, h->maxSG);
281
282#ifdef CONFIG_CISS_SCSI_TAPE
283	cciss_seq_tape_report(seq, h->ctlr);
284#endif /* CONFIG_CISS_SCSI_TAPE */
285}
286
287static void *cciss_seq_start(struct seq_file *seq, loff_t *pos)
288{
289	ctlr_info_t *h = seq->private;
290	unsigned ctlr = h->ctlr;
291	unsigned long flags;
292
293	/* prevent displaying bogus info during configuration
294	 * or deconfiguration of a logical volume
295	 */
296	spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
297	if (h->busy_configuring) {
298		spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
299		return ERR_PTR(-EBUSY);
300	}
301	h->busy_configuring = 1;
302	spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
303
304	if (*pos == 0)
305		cciss_seq_show_header(seq);
306
307	return pos;
308}
309
310static int cciss_seq_show(struct seq_file *seq, void *v)
311{
312	sector_t vol_sz, vol_sz_frac;
313	ctlr_info_t *h = seq->private;
314	unsigned ctlr = h->ctlr;
315	loff_t *pos = v;
316	drive_info_struct *drv = &h->drv[*pos];
317
318	if (*pos > h->highest_lun)
319		return 0;
320
321	if (drv->heads == 0)
322		return 0;
323
324	vol_sz = drv->nr_blocks;
325	vol_sz_frac = sector_div(vol_sz, ENG_GIG_FACTOR);
326	vol_sz_frac *= 100;
327	sector_div(vol_sz_frac, ENG_GIG_FACTOR);
328
329	if (drv->raid_level > 5)
330		drv->raid_level = RAID_UNKNOWN;
331	seq_printf(seq, "cciss/c%dd%d:"
332			"\t%4u.%02uGB\tRAID %s\n",
333			ctlr, (int) *pos, (int)vol_sz, (int)vol_sz_frac,
334			raid_label[drv->raid_level]);
335	return 0;
336}
337
338static void *cciss_seq_next(struct seq_file *seq, void *v, loff_t *pos)
339{
340	ctlr_info_t *h = seq->private;
341
342	if (*pos > h->highest_lun)
343		return NULL;
344	*pos += 1;
345
346	return pos;
347}
348
349static void cciss_seq_stop(struct seq_file *seq, void *v)
350{
351	ctlr_info_t *h = seq->private;
352
353	/* Only reset h->busy_configuring if we succeeded in setting
354	 * it during cciss_seq_start. */
355	if (v == ERR_PTR(-EBUSY))
356		return;
357
358	h->busy_configuring = 0;
359}
360
361static struct seq_operations cciss_seq_ops = {
362	.start = cciss_seq_start,
363	.show  = cciss_seq_show,
364	.next  = cciss_seq_next,
365	.stop  = cciss_seq_stop,
366};
367
368static int cciss_seq_open(struct inode *inode, struct file *file)
369{
370	int ret = seq_open(file, &cciss_seq_ops);
371	struct seq_file *seq = file->private_data;
372
373	if (!ret)
374		seq->private = PDE(inode)->data;
375
376	return ret;
377}
378
379static ssize_t
380cciss_proc_write(struct file *file, const char __user *buf,
381		 size_t length, loff_t *ppos)
382{
383	int err;
384	char *buffer;
385
386#ifndef CONFIG_CISS_SCSI_TAPE
387	return -EINVAL;
388#endif
389
390	if (!buf || length > PAGE_SIZE - 1)
391		return -EINVAL;
392
393	buffer = (char *)__get_free_page(GFP_KERNEL);
394	if (!buffer)
395		return -ENOMEM;
396
397	err = -EFAULT;
398	if (copy_from_user(buffer, buf, length))
399		goto out;
400	buffer[length] = '\0';
401
402#ifdef CONFIG_CISS_SCSI_TAPE
403	if (strncmp(ENGAGE_SCSI, buffer, sizeof ENGAGE_SCSI - 1) == 0) {
404		struct seq_file *seq = file->private_data;
405		ctlr_info_t *h = seq->private;
406		int rc;
407
408		rc = cciss_engage_scsi(h->ctlr);
409		if (rc != 0)
410			err = -rc;
411		else
412			err = length;
413	} else
414#endif /* CONFIG_CISS_SCSI_TAPE */
415		err = -EINVAL;
416	/* might be nice to have "disengage" too, but it's not
417	   safely possible. (only 1 module use count, lock issues.) */
418
419out:
420	free_page((unsigned long)buffer);
421	return err;
422}
423
424static struct file_operations cciss_proc_fops = {
425	.owner	 = THIS_MODULE,
426	.open    = cciss_seq_open,
427	.read    = seq_read,
428	.llseek  = seq_lseek,
429	.release = seq_release,
430	.write	 = cciss_proc_write,
431};
432
433static void __devinit cciss_procinit(int i)
434{
435	struct proc_dir_entry *pde;
436
437	if (proc_cciss == NULL)
438		proc_cciss = proc_mkdir("driver/cciss", NULL);
439	if (!proc_cciss)
440		return;
441	pde = proc_create_data(hba[i]->devname, S_IWUSR | S_IRUSR | S_IRGRP |
442					S_IROTH, proc_cciss,
443					&cciss_proc_fops, hba[i]);
444}
445#endif				/* CONFIG_PROC_FS */
446
447/*
448 * For operations that cannot sleep, a command block is allocated at init,
449 * and managed by cmd_alloc() and cmd_free() using a simple bitmap to track
450 * which ones are free or in use.  For operations that can wait for kmalloc
451 * to possible sleep, this routine can be called with get_from_pool set to 0.
452 * cmd_free() MUST be called with a got_from_pool set to 0 if cmd_alloc was.
453 */
454static CommandList_struct *cmd_alloc(ctlr_info_t *h, int get_from_pool)
455{
456	CommandList_struct *c;
457	int i;
458	u64bit temp64;
459	dma_addr_t cmd_dma_handle, err_dma_handle;
460
461	if (!get_from_pool) {
462		c = (CommandList_struct *) pci_alloc_consistent(h->pdev,
463			sizeof(CommandList_struct), &cmd_dma_handle);
464		if (c == NULL)
465			return NULL;
466		memset(c, 0, sizeof(CommandList_struct));
467
468		c->cmdindex = -1;
469
470		c->err_info = (ErrorInfo_struct *)
471		    pci_alloc_consistent(h->pdev, sizeof(ErrorInfo_struct),
472			    &err_dma_handle);
473
474		if (c->err_info == NULL) {
475			pci_free_consistent(h->pdev,
476				sizeof(CommandList_struct), c, cmd_dma_handle);
477			return NULL;
478		}
479		memset(c->err_info, 0, sizeof(ErrorInfo_struct));
480	} else {		/* get it out of the controllers pool */
481
482		do {
483			i = find_first_zero_bit(h->cmd_pool_bits, h->nr_cmds);
484			if (i == h->nr_cmds)
485				return NULL;
486		} while (test_and_set_bit
487			 (i & (BITS_PER_LONG - 1),
488			  h->cmd_pool_bits + (i / BITS_PER_LONG)) != 0);
489#ifdef CCISS_DEBUG
490		printk(KERN_DEBUG "cciss: using command buffer %d\n", i);
491#endif
492		c = h->cmd_pool + i;
493		memset(c, 0, sizeof(CommandList_struct));
494		cmd_dma_handle = h->cmd_pool_dhandle
495		    + i * sizeof(CommandList_struct);
496		c->err_info = h->errinfo_pool + i;
497		memset(c->err_info, 0, sizeof(ErrorInfo_struct));
498		err_dma_handle = h->errinfo_pool_dhandle
499		    + i * sizeof(ErrorInfo_struct);
500		h->nr_allocs++;
501
502		c->cmdindex = i;
503	}
504
505	c->busaddr = (__u32) cmd_dma_handle;
506	temp64.val = (__u64) err_dma_handle;
507	c->ErrDesc.Addr.lower = temp64.val32.lower;
508	c->ErrDesc.Addr.upper = temp64.val32.upper;
509	c->ErrDesc.Len = sizeof(ErrorInfo_struct);
510
511	c->ctlr = h->ctlr;
512	return c;
513}
514
515/*
516 * Frees a command block that was previously allocated with cmd_alloc().
517 */
518static void cmd_free(ctlr_info_t *h, CommandList_struct *c, int got_from_pool)
519{
520	int i;
521	u64bit temp64;
522
523	if (!got_from_pool) {
524		temp64.val32.lower = c->ErrDesc.Addr.lower;
525		temp64.val32.upper = c->ErrDesc.Addr.upper;
526		pci_free_consistent(h->pdev, sizeof(ErrorInfo_struct),
527				    c->err_info, (dma_addr_t) temp64.val);
528		pci_free_consistent(h->pdev, sizeof(CommandList_struct),
529				    c, (dma_addr_t) c->busaddr);
530	} else {
531		i = c - h->cmd_pool;
532		clear_bit(i & (BITS_PER_LONG - 1),
533			  h->cmd_pool_bits + (i / BITS_PER_LONG));
534		h->nr_frees++;
535	}
536}
537
538static inline ctlr_info_t *get_host(struct gendisk *disk)
539{
540	return disk->queue->queuedata;
541}
542
543static inline drive_info_struct *get_drv(struct gendisk *disk)
544{
545	return disk->private_data;
546}
547
548/*
549 * Open.  Make sure the device is really there.
550 */
551static int cciss_open(struct block_device *bdev, fmode_t mode)
552{
553	ctlr_info_t *host = get_host(bdev->bd_disk);
554	drive_info_struct *drv = get_drv(bdev->bd_disk);
555
556#ifdef CCISS_DEBUG
557	printk(KERN_DEBUG "cciss_open %s\n", bdev->bd_disk->disk_name);
558#endif				/* CCISS_DEBUG */
559
560	if (host->busy_initializing || drv->busy_configuring)
561		return -EBUSY;
562	/*
563	 * Root is allowed to open raw volume zero even if it's not configured
564	 * so array config can still work. Root is also allowed to open any
565	 * volume that has a LUN ID, so it can issue IOCTL to reread the
566	 * disk information.  I don't think I really like this
567	 * but I'm already using way to many device nodes to claim another one
568	 * for "raw controller".
569	 */
570	if (drv->heads == 0) {
571		if (MINOR(bdev->bd_dev) != 0) {	/* not node 0? */
572			/* if not node 0 make sure it is a partition = 0 */
573			if (MINOR(bdev->bd_dev) & 0x0f) {
574				return -ENXIO;
575				/* if it is, make sure we have a LUN ID */
576			} else if (drv->LunID == 0) {
577				return -ENXIO;
578			}
579		}
580		if (!capable(CAP_SYS_ADMIN))
581			return -EPERM;
582	}
583	drv->usage_count++;
584	host->usage_count++;
585	return 0;
586}
587
588/*
589 * Close.  Sync first.
590 */
591static int cciss_release(struct gendisk *disk, fmode_t mode)
592{
593	ctlr_info_t *host = get_host(disk);
594	drive_info_struct *drv = get_drv(disk);
595
596#ifdef CCISS_DEBUG
597	printk(KERN_DEBUG "cciss_release %s\n", disk->disk_name);
598#endif				/* CCISS_DEBUG */
599
600	drv->usage_count--;
601	host->usage_count--;
602	return 0;
603}
604
605#ifdef CONFIG_COMPAT
606
607static int do_ioctl(struct block_device *bdev, fmode_t mode,
608		    unsigned cmd, unsigned long arg)
609{
610	int ret;
611	lock_kernel();
612	ret = cciss_ioctl(bdev, mode, cmd, arg);
613	unlock_kernel();
614	return ret;
615}
616
617static int cciss_ioctl32_passthru(struct block_device *bdev, fmode_t mode,
618				  unsigned cmd, unsigned long arg);
619static int cciss_ioctl32_big_passthru(struct block_device *bdev, fmode_t mode,
620				      unsigned cmd, unsigned long arg);
621
622static int cciss_compat_ioctl(struct block_device *bdev, fmode_t mode,
623			      unsigned cmd, unsigned long arg)
624{
625	switch (cmd) {
626	case CCISS_GETPCIINFO:
627	case CCISS_GETINTINFO:
628	case CCISS_SETINTINFO:
629	case CCISS_GETNODENAME:
630	case CCISS_SETNODENAME:
631	case CCISS_GETHEARTBEAT:
632	case CCISS_GETBUSTYPES:
633	case CCISS_GETFIRMVER:
634	case CCISS_GETDRIVVER:
635	case CCISS_REVALIDVOLS:
636	case CCISS_DEREGDISK:
637	case CCISS_REGNEWDISK:
638	case CCISS_REGNEWD:
639	case CCISS_RESCANDISK:
640	case CCISS_GETLUNINFO:
641		return do_ioctl(bdev, mode, cmd, arg);
642
643	case CCISS_PASSTHRU32:
644		return cciss_ioctl32_passthru(bdev, mode, cmd, arg);
645	case CCISS_BIG_PASSTHRU32:
646		return cciss_ioctl32_big_passthru(bdev, mode, cmd, arg);
647
648	default:
649		return -ENOIOCTLCMD;
650	}
651}
652
653static int cciss_ioctl32_passthru(struct block_device *bdev, fmode_t mode,
654				  unsigned cmd, unsigned long arg)
655{
656	IOCTL32_Command_struct __user *arg32 =
657	    (IOCTL32_Command_struct __user *) arg;
658	IOCTL_Command_struct arg64;
659	IOCTL_Command_struct __user *p = compat_alloc_user_space(sizeof(arg64));
660	int err;
661	u32 cp;
662
663	err = 0;
664	err |=
665	    copy_from_user(&arg64.LUN_info, &arg32->LUN_info,
666			   sizeof(arg64.LUN_info));
667	err |=
668	    copy_from_user(&arg64.Request, &arg32->Request,
669			   sizeof(arg64.Request));
670	err |=
671	    copy_from_user(&arg64.error_info, &arg32->error_info,
672			   sizeof(arg64.error_info));
673	err |= get_user(arg64.buf_size, &arg32->buf_size);
674	err |= get_user(cp, &arg32->buf);
675	arg64.buf = compat_ptr(cp);
676	err |= copy_to_user(p, &arg64, sizeof(arg64));
677
678	if (err)
679		return -EFAULT;
680
681	err = do_ioctl(bdev, mode, CCISS_PASSTHRU, (unsigned long)p);
682	if (err)
683		return err;
684	err |=
685	    copy_in_user(&arg32->error_info, &p->error_info,
686			 sizeof(arg32->error_info));
687	if (err)
688		return -EFAULT;
689	return err;
690}
691
692static int cciss_ioctl32_big_passthru(struct block_device *bdev, fmode_t mode,
693				      unsigned cmd, unsigned long arg)
694{
695	BIG_IOCTL32_Command_struct __user *arg32 =
696	    (BIG_IOCTL32_Command_struct __user *) arg;
697	BIG_IOCTL_Command_struct arg64;
698	BIG_IOCTL_Command_struct __user *p =
699	    compat_alloc_user_space(sizeof(arg64));
700	int err;
701	u32 cp;
702
703	err = 0;
704	err |=
705	    copy_from_user(&arg64.LUN_info, &arg32->LUN_info,
706			   sizeof(arg64.LUN_info));
707	err |=
708	    copy_from_user(&arg64.Request, &arg32->Request,
709			   sizeof(arg64.Request));
710	err |=
711	    copy_from_user(&arg64.error_info, &arg32->error_info,
712			   sizeof(arg64.error_info));
713	err |= get_user(arg64.buf_size, &arg32->buf_size);
714	err |= get_user(arg64.malloc_size, &arg32->malloc_size);
715	err |= get_user(cp, &arg32->buf);
716	arg64.buf = compat_ptr(cp);
717	err |= copy_to_user(p, &arg64, sizeof(arg64));
718
719	if (err)
720		return -EFAULT;
721
722	err = do_ioctl(bdev, mode, CCISS_BIG_PASSTHRU, (unsigned long)p);
723	if (err)
724		return err;
725	err |=
726	    copy_in_user(&arg32->error_info, &p->error_info,
727			 sizeof(arg32->error_info));
728	if (err)
729		return -EFAULT;
730	return err;
731}
732#endif
733
734static int cciss_getgeo(struct block_device *bdev, struct hd_geometry *geo)
735{
736	drive_info_struct *drv = get_drv(bdev->bd_disk);
737
738	if (!drv->cylinders)
739		return -ENXIO;
740
741	geo->heads = drv->heads;
742	geo->sectors = drv->sectors;
743	geo->cylinders = drv->cylinders;
744	return 0;
745}
746
747/*
748 * ioctl
749 */
750static int cciss_ioctl(struct block_device *bdev, fmode_t mode,
751		       unsigned int cmd, unsigned long arg)
752{
753	struct gendisk *disk = bdev->bd_disk;
754	ctlr_info_t *host = get_host(disk);
755	drive_info_struct *drv = get_drv(disk);
756	int ctlr = host->ctlr;
757	void __user *argp = (void __user *)arg;
758
759#ifdef CCISS_DEBUG
760	printk(KERN_DEBUG "cciss_ioctl: Called with cmd=%x %lx\n", cmd, arg);
761#endif				/* CCISS_DEBUG */
762
763	switch (cmd) {
764	case CCISS_GETPCIINFO:
765		{
766			cciss_pci_info_struct pciinfo;
767
768			if (!arg)
769				return -EINVAL;
770			pciinfo.domain = pci_domain_nr(host->pdev->bus);
771			pciinfo.bus = host->pdev->bus->number;
772			pciinfo.dev_fn = host->pdev->devfn;
773			pciinfo.board_id = host->board_id;
774			if (copy_to_user
775			    (argp, &pciinfo, sizeof(cciss_pci_info_struct)))
776				return -EFAULT;
777			return 0;
778		}
779	case CCISS_GETINTINFO:
780		{
781			cciss_coalint_struct intinfo;
782			if (!arg)
783				return -EINVAL;
784			intinfo.delay =
785			    readl(&host->cfgtable->HostWrite.CoalIntDelay);
786			intinfo.count =
787			    readl(&host->cfgtable->HostWrite.CoalIntCount);
788			if (copy_to_user
789			    (argp, &intinfo, sizeof(cciss_coalint_struct)))
790				return -EFAULT;
791			return 0;
792		}
793	case CCISS_SETINTINFO:
794		{
795			cciss_coalint_struct intinfo;
796			unsigned long flags;
797			int i;
798
799			if (!arg)
800				return -EINVAL;
801			if (!capable(CAP_SYS_ADMIN))
802				return -EPERM;
803			if (copy_from_user
804			    (&intinfo, argp, sizeof(cciss_coalint_struct)))
805				return -EFAULT;
806			if ((intinfo.delay == 0) && (intinfo.count == 0))
807			{
808//                      printk("cciss_ioctl: delay and count cannot be 0\n");
809				return -EINVAL;
810			}
811			spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
812			/* Update the field, and then ring the doorbell */
813			writel(intinfo.delay,
814			       &(host->cfgtable->HostWrite.CoalIntDelay));
815			writel(intinfo.count,
816			       &(host->cfgtable->HostWrite.CoalIntCount));
817			writel(CFGTBL_ChangeReq, host->vaddr + SA5_DOORBELL);
818
819			for (i = 0; i < MAX_IOCTL_CONFIG_WAIT; i++) {
820				if (!(readl(host->vaddr + SA5_DOORBELL)
821				      & CFGTBL_ChangeReq))
822					break;
823				/* delay and try again */
824				udelay(1000);
825			}
826			spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
827			if (i >= MAX_IOCTL_CONFIG_WAIT)
828				return -EAGAIN;
829			return 0;
830		}
831	case CCISS_GETNODENAME:
832		{
833			NodeName_type NodeName;
834			int i;
835
836			if (!arg)
837				return -EINVAL;
838			for (i = 0; i < 16; i++)
839				NodeName[i] =
840				    readb(&host->cfgtable->ServerName[i]);
841			if (copy_to_user(argp, NodeName, sizeof(NodeName_type)))
842				return -EFAULT;
843			return 0;
844		}
845	case CCISS_SETNODENAME:
846		{
847			NodeName_type NodeName;
848			unsigned long flags;
849			int i;
850
851			if (!arg)
852				return -EINVAL;
853			if (!capable(CAP_SYS_ADMIN))
854				return -EPERM;
855
856			if (copy_from_user
857			    (NodeName, argp, sizeof(NodeName_type)))
858				return -EFAULT;
859
860			spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
861
862			/* Update the field, and then ring the doorbell */
863			for (i = 0; i < 16; i++)
864				writeb(NodeName[i],
865				       &host->cfgtable->ServerName[i]);
866
867			writel(CFGTBL_ChangeReq, host->vaddr + SA5_DOORBELL);
868
869			for (i = 0; i < MAX_IOCTL_CONFIG_WAIT; i++) {
870				if (!(readl(host->vaddr + SA5_DOORBELL)
871				      & CFGTBL_ChangeReq))
872					break;
873				/* delay and try again */
874				udelay(1000);
875			}
876			spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
877			if (i >= MAX_IOCTL_CONFIG_WAIT)
878				return -EAGAIN;
879			return 0;
880		}
881
882	case CCISS_GETHEARTBEAT:
883		{
884			Heartbeat_type heartbeat;
885
886			if (!arg)
887				return -EINVAL;
888			heartbeat = readl(&host->cfgtable->HeartBeat);
889			if (copy_to_user
890			    (argp, &heartbeat, sizeof(Heartbeat_type)))
891				return -EFAULT;
892			return 0;
893		}
894	case CCISS_GETBUSTYPES:
895		{
896			BusTypes_type BusTypes;
897
898			if (!arg)
899				return -EINVAL;
900			BusTypes = readl(&host->cfgtable->BusTypes);
901			if (copy_to_user
902			    (argp, &BusTypes, sizeof(BusTypes_type)))
903				return -EFAULT;
904			return 0;
905		}
906	case CCISS_GETFIRMVER:
907		{
908			FirmwareVer_type firmware;
909
910			if (!arg)
911				return -EINVAL;
912			memcpy(firmware, host->firm_ver, 4);
913
914			if (copy_to_user
915			    (argp, firmware, sizeof(FirmwareVer_type)))
916				return -EFAULT;
917			return 0;
918		}
919	case CCISS_GETDRIVVER:
920		{
921			DriverVer_type DriverVer = DRIVER_VERSION;
922
923			if (!arg)
924				return -EINVAL;
925
926			if (copy_to_user
927			    (argp, &DriverVer, sizeof(DriverVer_type)))
928				return -EFAULT;
929			return 0;
930		}
931
932	case CCISS_DEREGDISK:
933	case CCISS_REGNEWD:
934	case CCISS_REVALIDVOLS:
935		return rebuild_lun_table(host, 0);
936
937	case CCISS_GETLUNINFO:{
938			LogvolInfo_struct luninfo;
939
940			luninfo.LunID = drv->LunID;
941			luninfo.num_opens = drv->usage_count;
942			luninfo.num_parts = 0;
943			if (copy_to_user(argp, &luninfo,
944					 sizeof(LogvolInfo_struct)))
945				return -EFAULT;
946			return 0;
947		}
948	case CCISS_PASSTHRU:
949		{
950			IOCTL_Command_struct iocommand;
951			CommandList_struct *c;
952			char *buff = NULL;
953			u64bit temp64;
954			unsigned long flags;
955			DECLARE_COMPLETION_ONSTACK(wait);
956
957			if (!arg)
958				return -EINVAL;
959
960			if (!capable(CAP_SYS_RAWIO))
961				return -EPERM;
962
963			if (copy_from_user
964			    (&iocommand, argp, sizeof(IOCTL_Command_struct)))
965				return -EFAULT;
966			if ((iocommand.buf_size < 1) &&
967			    (iocommand.Request.Type.Direction != XFER_NONE)) {
968				return -EINVAL;
969			}
970#if 0				/* 'buf_size' member is 16-bits, and always smaller than kmalloc limit */
971			/* Check kmalloc limits */
972			if (iocommand.buf_size > 128000)
973				return -EINVAL;
974#endif
975			if (iocommand.buf_size > 0) {
976				buff = kmalloc(iocommand.buf_size, GFP_KERNEL);
977				if (buff == NULL)
978					return -EFAULT;
979			}
980			if (iocommand.Request.Type.Direction == XFER_WRITE) {
981				/* Copy the data into the buffer we created */
982				if (copy_from_user
983				    (buff, iocommand.buf, iocommand.buf_size)) {
984					kfree(buff);
985					return -EFAULT;
986				}
987			} else {
988				memset(buff, 0, iocommand.buf_size);
989			}
990			if ((c = cmd_alloc(host, 0)) == NULL) {
991				kfree(buff);
992				return -ENOMEM;
993			}
994			// Fill in the command type
995			c->cmd_type = CMD_IOCTL_PEND;
996			// Fill in Command Header
997			c->Header.ReplyQueue = 0;	// unused in simple mode
998			if (iocommand.buf_size > 0)	// buffer to fill
999			{
1000				c->Header.SGList = 1;
1001				c->Header.SGTotal = 1;
1002			} else	// no buffers to fill
1003			{
1004				c->Header.SGList = 0;
1005				c->Header.SGTotal = 0;
1006			}
1007			c->Header.LUN = iocommand.LUN_info;
1008			c->Header.Tag.lower = c->busaddr;	// use the kernel address the cmd block for tag
1009
1010			// Fill in Request block
1011			c->Request = iocommand.Request;
1012
1013			// Fill in the scatter gather information
1014			if (iocommand.buf_size > 0) {
1015				temp64.val = pci_map_single(host->pdev, buff,
1016					iocommand.buf_size,
1017					PCI_DMA_BIDIRECTIONAL);
1018				c->SG[0].Addr.lower = temp64.val32.lower;
1019				c->SG[0].Addr.upper = temp64.val32.upper;
1020				c->SG[0].Len = iocommand.buf_size;
1021				c->SG[0].Ext = 0;	// we are not chaining
1022			}
1023			c->waiting = &wait;
1024
1025			/* Put the request on the tail of the request queue */
1026			spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
1027			addQ(&host->reqQ, c);
1028			host->Qdepth++;
1029			start_io(host);
1030			spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
1031
1032			wait_for_completion(&wait);
1033
1034			/* unlock the buffers from DMA */
1035			temp64.val32.lower = c->SG[0].Addr.lower;
1036			temp64.val32.upper = c->SG[0].Addr.upper;
1037			pci_unmap_single(host->pdev, (dma_addr_t) temp64.val,
1038					 iocommand.buf_size,
1039					 PCI_DMA_BIDIRECTIONAL);
1040
1041			/* Copy the error information out */
1042			iocommand.error_info = *(c->err_info);
1043			if (copy_to_user
1044			    (argp, &iocommand, sizeof(IOCTL_Command_struct))) {
1045				kfree(buff);
1046				cmd_free(host, c, 0);
1047				return -EFAULT;
1048			}
1049
1050			if (iocommand.Request.Type.Direction == XFER_READ) {
1051				/* Copy the data out of the buffer we created */
1052				if (copy_to_user
1053				    (iocommand.buf, buff, iocommand.buf_size)) {
1054					kfree(buff);
1055					cmd_free(host, c, 0);
1056					return -EFAULT;
1057				}
1058			}
1059			kfree(buff);
1060			cmd_free(host, c, 0);
1061			return 0;
1062		}
1063	case CCISS_BIG_PASSTHRU:{
1064			BIG_IOCTL_Command_struct *ioc;
1065			CommandList_struct *c;
1066			unsigned char **buff = NULL;
1067			int *buff_size = NULL;
1068			u64bit temp64;
1069			unsigned long flags;
1070			BYTE sg_used = 0;
1071			int status = 0;
1072			int i;
1073			DECLARE_COMPLETION_ONSTACK(wait);
1074			__u32 left;
1075			__u32 sz;
1076			BYTE __user *data_ptr;
1077
1078			if (!arg)
1079				return -EINVAL;
1080			if (!capable(CAP_SYS_RAWIO))
1081				return -EPERM;
1082			ioc = (BIG_IOCTL_Command_struct *)
1083			    kmalloc(sizeof(*ioc), GFP_KERNEL);
1084			if (!ioc) {
1085				status = -ENOMEM;
1086				goto cleanup1;
1087			}
1088			if (copy_from_user(ioc, argp, sizeof(*ioc))) {
1089				status = -EFAULT;
1090				goto cleanup1;
1091			}
1092			if ((ioc->buf_size < 1) &&
1093			    (ioc->Request.Type.Direction != XFER_NONE)) {
1094				status = -EINVAL;
1095				goto cleanup1;
1096			}
1097			/* Check kmalloc limits  using all SGs */
1098			if (ioc->malloc_size > MAX_KMALLOC_SIZE) {
1099				status = -EINVAL;
1100				goto cleanup1;
1101			}
1102			if (ioc->buf_size > ioc->malloc_size * MAXSGENTRIES) {
1103				status = -EINVAL;
1104				goto cleanup1;
1105			}
1106			buff =
1107			    kzalloc(MAXSGENTRIES * sizeof(char *), GFP_KERNEL);
1108			if (!buff) {
1109				status = -ENOMEM;
1110				goto cleanup1;
1111			}
1112			buff_size = kmalloc(MAXSGENTRIES * sizeof(int),
1113						   GFP_KERNEL);
1114			if (!buff_size) {
1115				status = -ENOMEM;
1116				goto cleanup1;
1117			}
1118			left = ioc->buf_size;
1119			data_ptr = ioc->buf;
1120			while (left) {
1121				sz = (left >
1122				      ioc->malloc_size) ? ioc->
1123				    malloc_size : left;
1124				buff_size[sg_used] = sz;
1125				buff[sg_used] = kmalloc(sz, GFP_KERNEL);
1126				if (buff[sg_used] == NULL) {
1127					status = -ENOMEM;
1128					goto cleanup1;
1129				}
1130				if (ioc->Request.Type.Direction == XFER_WRITE) {
1131					if (copy_from_user
1132					    (buff[sg_used], data_ptr, sz)) {
1133						status = -EFAULT;
1134						goto cleanup1;
1135					}
1136				} else {
1137					memset(buff[sg_used], 0, sz);
1138				}
1139				left -= sz;
1140				data_ptr += sz;
1141				sg_used++;
1142			}
1143			if ((c = cmd_alloc(host, 0)) == NULL) {
1144				status = -ENOMEM;
1145				goto cleanup1;
1146			}
1147			c->cmd_type = CMD_IOCTL_PEND;
1148			c->Header.ReplyQueue = 0;
1149
1150			if (ioc->buf_size > 0) {
1151				c->Header.SGList = sg_used;
1152				c->Header.SGTotal = sg_used;
1153			} else {
1154				c->Header.SGList = 0;
1155				c->Header.SGTotal = 0;
1156			}
1157			c->Header.LUN = ioc->LUN_info;
1158			c->Header.Tag.lower = c->busaddr;
1159
1160			c->Request = ioc->Request;
1161			if (ioc->buf_size > 0) {
1162				int i;
1163				for (i = 0; i < sg_used; i++) {
1164					temp64.val =
1165					    pci_map_single(host->pdev, buff[i],
1166						    buff_size[i],
1167						    PCI_DMA_BIDIRECTIONAL);
1168					c->SG[i].Addr.lower =
1169					    temp64.val32.lower;
1170					c->SG[i].Addr.upper =
1171					    temp64.val32.upper;
1172					c->SG[i].Len = buff_size[i];
1173					c->SG[i].Ext = 0;	/* we are not chaining */
1174				}
1175			}
1176			c->waiting = &wait;
1177			/* Put the request on the tail of the request queue */
1178			spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
1179			addQ(&host->reqQ, c);
1180			host->Qdepth++;
1181			start_io(host);
1182			spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
1183			wait_for_completion(&wait);
1184			/* unlock the buffers from DMA */
1185			for (i = 0; i < sg_used; i++) {
1186				temp64.val32.lower = c->SG[i].Addr.lower;
1187				temp64.val32.upper = c->SG[i].Addr.upper;
1188				pci_unmap_single(host->pdev,
1189					(dma_addr_t) temp64.val, buff_size[i],
1190					PCI_DMA_BIDIRECTIONAL);
1191			}
1192			/* Copy the error information out */
1193			ioc->error_info = *(c->err_info);
1194			if (copy_to_user(argp, ioc, sizeof(*ioc))) {
1195				cmd_free(host, c, 0);
1196				status = -EFAULT;
1197				goto cleanup1;
1198			}
1199			if (ioc->Request.Type.Direction == XFER_READ) {
1200				/* Copy the data out of the buffer we created */
1201				BYTE __user *ptr = ioc->buf;
1202				for (i = 0; i < sg_used; i++) {
1203					if (copy_to_user
1204					    (ptr, buff[i], buff_size[i])) {
1205						cmd_free(host, c, 0);
1206						status = -EFAULT;
1207						goto cleanup1;
1208					}
1209					ptr += buff_size[i];
1210				}
1211			}
1212			cmd_free(host, c, 0);
1213			status = 0;
1214		      cleanup1:
1215			if (buff) {
1216				for (i = 0; i < sg_used; i++)
1217					kfree(buff[i]);
1218				kfree(buff);
1219			}
1220			kfree(buff_size);
1221			kfree(ioc);
1222			return status;
1223		}
1224
1225	/* scsi_cmd_ioctl handles these, below, though some are not */
1226	/* very meaningful for cciss.  SG_IO is the main one people want. */
1227
1228	case SG_GET_VERSION_NUM:
1229	case SG_SET_TIMEOUT:
1230	case SG_GET_TIMEOUT:
1231	case SG_GET_RESERVED_SIZE:
1232	case SG_SET_RESERVED_SIZE:
1233	case SG_EMULATED_HOST:
1234	case SG_IO:
1235	case SCSI_IOCTL_SEND_COMMAND:
1236		return scsi_cmd_ioctl(disk->queue, disk, mode, cmd, argp);
1237
1238	/* scsi_cmd_ioctl would normally handle these, below, but */
1239	/* they aren't a good fit for cciss, as CD-ROMs are */
1240	/* not supported, and we don't have any bus/target/lun */
1241	/* which we present to the kernel. */
1242
1243	case CDROM_SEND_PACKET:
1244	case CDROMCLOSETRAY:
1245	case CDROMEJECT:
1246	case SCSI_IOCTL_GET_IDLUN:
1247	case SCSI_IOCTL_GET_BUS_NUMBER:
1248	default:
1249		return -ENOTTY;
1250	}
1251}
1252
1253static void cciss_check_queues(ctlr_info_t *h)
1254{
1255	int start_queue = h->next_to_run;
1256	int i;
1257
1258	/* check to see if we have maxed out the number of commands that can
1259	 * be placed on the queue.  If so then exit.  We do this check here
1260	 * in case the interrupt we serviced was from an ioctl and did not
1261	 * free any new commands.
1262	 */
1263	if ((find_first_zero_bit(h->cmd_pool_bits, h->nr_cmds)) == h->nr_cmds)
1264		return;
1265
1266	/* We have room on the queue for more commands.  Now we need to queue
1267	 * them up.  We will also keep track of the next queue to run so
1268	 * that every queue gets a chance to be started first.
1269	 */
1270	for (i = 0; i < h->highest_lun + 1; i++) {
1271		int curr_queue = (start_queue + i) % (h->highest_lun + 1);
1272		/* make sure the disk has been added and the drive is real
1273		 * because this can be called from the middle of init_one.
1274		 */
1275		if (!(h->drv[curr_queue].queue) || !(h->drv[curr_queue].heads))
1276			continue;
1277		blk_start_queue(h->gendisk[curr_queue]->queue);
1278
1279		/* check to see if we have maxed out the number of commands
1280		 * that can be placed on the queue.
1281		 */
1282		if ((find_first_zero_bit(h->cmd_pool_bits, h->nr_cmds)) == h->nr_cmds) {
1283			if (curr_queue == start_queue) {
1284				h->next_to_run =
1285				    (start_queue + 1) % (h->highest_lun + 1);
1286				break;
1287			} else {
1288				h->next_to_run = curr_queue;
1289				break;
1290			}
1291		}
1292	}
1293}
1294
1295static void cciss_softirq_done(struct request *rq)
1296{
1297	CommandList_struct *cmd = rq->completion_data;
1298	ctlr_info_t *h = hba[cmd->ctlr];
1299	unsigned long flags;
1300	u64bit temp64;
1301	int i, ddir;
1302
1303	if (cmd->Request.Type.Direction == XFER_READ)
1304		ddir = PCI_DMA_FROMDEVICE;
1305	else
1306		ddir = PCI_DMA_TODEVICE;
1307
1308	/* command did not need to be retried */
1309	/* unmap the DMA mapping for all the scatter gather elements */
1310	for (i = 0; i < cmd->Header.SGList; i++) {
1311		temp64.val32.lower = cmd->SG[i].Addr.lower;
1312		temp64.val32.upper = cmd->SG[i].Addr.upper;
1313		pci_unmap_page(h->pdev, temp64.val, cmd->SG[i].Len, ddir);
1314	}
1315
1316#ifdef CCISS_DEBUG
1317	printk("Done with %p\n", rq);
1318#endif				/* CCISS_DEBUG */
1319
1320	if (blk_end_request(rq, (rq->errors == 0) ? 0 : -EIO, blk_rq_bytes(rq)))
1321		BUG();
1322
1323	spin_lock_irqsave(&h->lock, flags);
1324	cmd_free(h, cmd, 1);
1325	cciss_check_queues(h);
1326	spin_unlock_irqrestore(&h->lock, flags);
1327}
1328
1329/* This function gets the serial number of a logical drive via
1330 * inquiry page 0x83.  Serial no. is 16 bytes.  If the serial
1331 * number cannot be had, for whatever reason, 16 bytes of 0xff
1332 * are returned instead.
1333 */
1334static void cciss_get_serial_no(int ctlr, int logvol, int withirq,
1335				unsigned char *serial_no, int buflen)
1336{
1337#define PAGE_83_INQ_BYTES 64
1338	int rc;
1339	unsigned char *buf;
1340
1341	if (buflen > 16)
1342		buflen = 16;
1343	memset(serial_no, 0xff, buflen);
1344	buf = kzalloc(PAGE_83_INQ_BYTES, GFP_KERNEL);
1345	if (!buf)
1346		return;
1347	memset(serial_no, 0, buflen);
1348	if (withirq)
1349		rc = sendcmd_withirq(CISS_INQUIRY, ctlr, buf,
1350			PAGE_83_INQ_BYTES, 1, logvol, 0x83, TYPE_CMD);
1351	else
1352		rc = sendcmd(CISS_INQUIRY, ctlr, buf,
1353			PAGE_83_INQ_BYTES, 1, logvol, 0x83, NULL, TYPE_CMD);
1354	if (rc == IO_OK)
1355		memcpy(serial_no, &buf[8], buflen);
1356	kfree(buf);
1357	return;
1358}
1359
1360static void cciss_add_disk(ctlr_info_t *h, struct gendisk *disk,
1361				int drv_index)
1362{
1363	disk->queue = blk_init_queue(do_cciss_request, &h->lock);
1364	sprintf(disk->disk_name, "cciss/c%dd%d", h->ctlr, drv_index);
1365	disk->major = h->major;
1366	disk->first_minor = drv_index << NWD_SHIFT;
1367	disk->fops = &cciss_fops;
1368	disk->private_data = &h->drv[drv_index];
1369
1370	/* Set up queue information */
1371	blk_queue_bounce_limit(disk->queue, h->pdev->dma_mask);
1372
1373	/* This is a hardware imposed limit. */
1374	blk_queue_max_hw_segments(disk->queue, MAXSGENTRIES);
1375
1376	/* This is a limit in the driver and could be eliminated. */
1377	blk_queue_max_phys_segments(disk->queue, MAXSGENTRIES);
1378
1379	blk_queue_max_sectors(disk->queue, h->cciss_max_sectors);
1380
1381	blk_queue_softirq_done(disk->queue, cciss_softirq_done);
1382
1383	disk->queue->queuedata = h;
1384
1385	blk_queue_hardsect_size(disk->queue,
1386				h->drv[drv_index].block_size);
1387
1388	/* Make sure all queue data is written out before */
1389	/* setting h->drv[drv_index].queue, as setting this */
1390	/* allows the interrupt handler to start the queue */
1391	wmb();
1392	h->drv[drv_index].queue = disk->queue;
1393	add_disk(disk);
1394}
1395
1396/* This function will check the usage_count of the drive to be updated/added.
1397 * If the usage_count is zero and it is a heretofore unknown drive, or,
1398 * the drive's capacity, geometry, or serial number has changed,
1399 * then the drive information will be updated and the disk will be
1400 * re-registered with the kernel.  If these conditions don't hold,
1401 * then it will be left alone for the next reboot.  The exception to this
1402 * is disk 0 which will always be left registered with the kernel since it
1403 * is also the controller node.  Any changes to disk 0 will show up on
1404 * the next reboot.
1405 */
1406static void cciss_update_drive_info(int ctlr, int drv_index, int first_time)
1407{
1408	ctlr_info_t *h = hba[ctlr];
1409	struct gendisk *disk;
1410	InquiryData_struct *inq_buff = NULL;
1411	unsigned int block_size;
1412	sector_t total_size;
1413	unsigned long flags = 0;
1414	int ret = 0;
1415	drive_info_struct *drvinfo;
1416	int was_only_controller_node;
1417
1418	/* Get information about the disk and modify the driver structure */
1419	inq_buff = kmalloc(sizeof(InquiryData_struct), GFP_KERNEL);
1420	drvinfo = kmalloc(sizeof(*drvinfo), GFP_KERNEL);
1421	if (inq_buff == NULL || drvinfo == NULL)
1422		goto mem_msg;
1423
1424	/* See if we're trying to update the "controller node"
1425	 * this will happen the when the first logical drive gets
1426	 * created by ACU.
1427	 */
1428	was_only_controller_node = (drv_index == 0 &&
1429				h->drv[0].raid_level == -1);
1430
1431	/* testing to see if 16-byte CDBs are already being used */
1432	if (h->cciss_read == CCISS_READ_16) {
1433		cciss_read_capacity_16(h->ctlr, drv_index, 1,
1434			&total_size, &block_size);
1435
1436	} else {
1437		cciss_read_capacity(ctlr, drv_index, 1,
1438				    &total_size, &block_size);
1439
1440		/* if read_capacity returns all F's this volume is >2TB */
1441		/* in size so we switch to 16-byte CDB's for all */
1442		/* read/write ops */
1443		if (total_size == 0xFFFFFFFFULL) {
1444			cciss_read_capacity_16(ctlr, drv_index, 1,
1445			&total_size, &block_size);
1446			h->cciss_read = CCISS_READ_16;
1447			h->cciss_write = CCISS_WRITE_16;
1448		} else {
1449			h->cciss_read = CCISS_READ_10;
1450			h->cciss_write = CCISS_WRITE_10;
1451		}
1452	}
1453
1454	cciss_geometry_inquiry(ctlr, drv_index, 1, total_size, block_size,
1455			       inq_buff, drvinfo);
1456	drvinfo->block_size = block_size;
1457	drvinfo->nr_blocks = total_size + 1;
1458
1459	cciss_get_serial_no(ctlr, drv_index, 1, drvinfo->serial_no,
1460			sizeof(drvinfo->serial_no));
1461
1462	/* Is it the same disk we already know, and nothing's changed? */
1463	if (h->drv[drv_index].raid_level != -1 &&
1464		((memcmp(drvinfo->serial_no,
1465				h->drv[drv_index].serial_no, 16) == 0) &&
1466		drvinfo->block_size == h->drv[drv_index].block_size &&
1467		drvinfo->nr_blocks == h->drv[drv_index].nr_blocks &&
1468		drvinfo->heads == h->drv[drv_index].heads &&
1469		drvinfo->sectors == h->drv[drv_index].sectors &&
1470		drvinfo->cylinders == h->drv[drv_index].cylinders))
1471			/* The disk is unchanged, nothing to update */
1472			goto freeret;
1473
1474	/* If we get here it's not the same disk, or something's changed,
1475	 * so we need to * deregister it, and re-register it, if it's not
1476	 * in use.
1477	 * If the disk already exists then deregister it before proceeding
1478	 * (unless it's the first disk (for the controller node).
1479	 */
1480	if (h->drv[drv_index].raid_level != -1 && drv_index != 0) {
1481		printk(KERN_WARNING "disk %d has changed.\n", drv_index);
1482		spin_lock_irqsave(CCISS_LOCK(h->ctlr), flags);
1483		h->drv[drv_index].busy_configuring = 1;
1484		spin_unlock_irqrestore(CCISS_LOCK(h->ctlr), flags);
1485
1486		/* deregister_disk sets h->drv[drv_index].queue = NULL
1487		 * which keeps the interrupt handler from starting
1488		 * the queue.
1489		 */
1490		ret = deregister_disk(h->gendisk[drv_index],
1491				      &h->drv[drv_index], 0);
1492		h->drv[drv_index].busy_configuring = 0;
1493	}
1494
1495	/* If the disk is in use return */
1496	if (ret)
1497		goto freeret;
1498
1499	/* Save the new information from cciss_geometry_inquiry
1500	 * and serial number inquiry.
1501	 */
1502	h->drv[drv_index].block_size = drvinfo->block_size;
1503	h->drv[drv_index].nr_blocks = drvinfo->nr_blocks;
1504	h->drv[drv_index].heads = drvinfo->heads;
1505	h->drv[drv_index].sectors = drvinfo->sectors;
1506	h->drv[drv_index].cylinders = drvinfo->cylinders;
1507	h->drv[drv_index].raid_level = drvinfo->raid_level;
1508	memcpy(h->drv[drv_index].serial_no, drvinfo->serial_no, 16);
1509
1510	++h->num_luns;
1511	disk = h->gendisk[drv_index];
1512	set_capacity(disk, h->drv[drv_index].nr_blocks);
1513
1514	/* If it's not disk 0 (drv_index != 0)
1515	 * or if it was disk 0, but there was previously
1516	 * no actual corresponding configured logical drive
1517	 * (raid_leve == -1) then we want to update the
1518	 * logical drive's information.
1519	 */
1520	if (drv_index || first_time)
1521		cciss_add_disk(h, disk, drv_index);
1522
1523freeret:
1524	kfree(inq_buff);
1525	kfree(drvinfo);
1526	return;
1527mem_msg:
1528	printk(KERN_ERR "cciss: out of memory\n");
1529	goto freeret;
1530}
1531
1532/* This function will find the first index of the controllers drive array
1533 * that has a -1 for the raid_level and will return that index.  This is
1534 * where new drives will be added.  If the index to be returned is greater
1535 * than the highest_lun index for the controller then highest_lun is set
1536 * to this new index.  If there are no available indexes then -1 is returned.
1537 * "controller_node" is used to know if this is a real logical drive, or just
1538 * the controller node, which determines if this counts towards highest_lun.
1539 */
1540static int cciss_find_free_drive_index(int ctlr, int controller_node)
1541{
1542	int i;
1543
1544	for (i = 0; i < CISS_MAX_LUN; i++) {
1545		if (hba[ctlr]->drv[i].raid_level == -1) {
1546			if (i > hba[ctlr]->highest_lun)
1547				if (!controller_node)
1548					hba[ctlr]->highest_lun = i;
1549			return i;
1550		}
1551	}
1552	return -1;
1553}
1554
1555/* cciss_add_gendisk finds a free hba[]->drv structure
1556 * and allocates a gendisk if needed, and sets the lunid
1557 * in the drvinfo structure.   It returns the index into
1558 * the ->drv[] array, or -1 if none are free.
1559 * is_controller_node indicates whether highest_lun should
1560 * count this disk, or if it's only being added to provide
1561 * a means to talk to the controller in case no logical
1562 * drives have yet been configured.
1563 */
1564static int cciss_add_gendisk(ctlr_info_t *h, __u32 lunid, int controller_node)
1565{
1566	int drv_index;
1567
1568	drv_index = cciss_find_free_drive_index(h->ctlr, controller_node);
1569	if (drv_index == -1)
1570		return -1;
1571	/*Check if the gendisk needs to be allocated */
1572	if (!h->gendisk[drv_index]) {
1573		h->gendisk[drv_index] =
1574			alloc_disk(1 << NWD_SHIFT);
1575		if (!h->gendisk[drv_index]) {
1576			printk(KERN_ERR "cciss%d: could not "
1577				"allocate a new disk %d\n",
1578				h->ctlr, drv_index);
1579			return -1;
1580		}
1581	}
1582	h->drv[drv_index].LunID = lunid;
1583
1584	/* Don't need to mark this busy because nobody */
1585	/* else knows about this disk yet to contend */
1586	/* for access to it. */
1587	h->drv[drv_index].busy_configuring = 0;
1588	wmb();
1589	return drv_index;
1590}
1591
1592/* This is for the special case of a controller which
1593 * has no logical drives.  In this case, we still need
1594 * to register a disk so the controller can be accessed
1595 * by the Array Config Utility.
1596 */
1597static void cciss_add_controller_node(ctlr_info_t *h)
1598{
1599	struct gendisk *disk;
1600	int drv_index;
1601
1602	if (h->gendisk[0] != NULL) /* already did this? Then bail. */
1603		return;
1604
1605	drv_index = cciss_add_gendisk(h, 0, 1);
1606	if (drv_index == -1) {
1607		printk(KERN_WARNING "cciss%d: could not "
1608			"add disk 0.\n", h->ctlr);
1609		return;
1610	}
1611	h->drv[drv_index].block_size = 512;
1612	h->drv[drv_index].nr_blocks = 0;
1613	h->drv[drv_index].heads = 0;
1614	h->drv[drv_index].sectors = 0;
1615	h->drv[drv_index].cylinders = 0;
1616	h->drv[drv_index].raid_level = -1;
1617	memset(h->drv[drv_index].serial_no, 0, 16);
1618	disk = h->gendisk[drv_index];
1619	cciss_add_disk(h, disk, drv_index);
1620}
1621
1622/* This function will add and remove logical drives from the Logical
1623 * drive array of the controller and maintain persistency of ordering
1624 * so that mount points are preserved until the next reboot.  This allows
1625 * for the removal of logical drives in the middle of the drive array
1626 * without a re-ordering of those drives.
1627 * INPUT
1628 * h		= The controller to perform the operations on
1629 */
1630static int rebuild_lun_table(ctlr_info_t *h, int first_time)
1631{
1632	int ctlr = h->ctlr;
1633	int num_luns;
1634	ReportLunData_struct *ld_buff = NULL;
1635	int return_code;
1636	int listlength = 0;
1637	int i;
1638	int drv_found;
1639	int drv_index = 0;
1640	__u32 lunid = 0;
1641	unsigned long flags;
1642
1643	if (!capable(CAP_SYS_RAWIO))
1644		return -EPERM;
1645
1646	/* Set busy_configuring flag for this operation */
1647	spin_lock_irqsave(CCISS_LOCK(h->ctlr), flags);
1648	if (h->busy_configuring) {
1649		spin_unlock_irqrestore(CCISS_LOCK(h->ctlr), flags);
1650		return -EBUSY;
1651	}
1652	h->busy_configuring = 1;
1653	spin_unlock_irqrestore(CCISS_LOCK(h->ctlr), flags);
1654
1655	ld_buff = kzalloc(sizeof(ReportLunData_struct), GFP_KERNEL);
1656	if (ld_buff == NULL)
1657		goto mem_msg;
1658
1659	return_code = sendcmd_withirq(CISS_REPORT_LOG, ctlr, ld_buff,
1660				      sizeof(ReportLunData_struct), 0,
1661				      0, 0, TYPE_CMD);
1662
1663	if (return_code == IO_OK)
1664		listlength = be32_to_cpu(*(__be32 *) ld_buff->LUNListLength);
1665	else {	/* reading number of logical volumes failed */
1666		printk(KERN_WARNING "cciss: report logical volume"
1667		       " command failed\n");
1668		listlength = 0;
1669		goto freeret;
1670	}
1671
1672	num_luns = listlength / 8;	/* 8 bytes per entry */
1673	if (num_luns > CISS_MAX_LUN) {
1674		num_luns = CISS_MAX_LUN;
1675		printk(KERN_WARNING "cciss: more luns configured"
1676		       " on controller than can be handled by"
1677		       " this driver.\n");
1678	}
1679
1680	if (num_luns == 0)
1681		cciss_add_controller_node(h);
1682
1683	/* Compare controller drive array to driver's drive array
1684	 * to see if any drives are missing on the controller due
1685	 * to action of Array Config Utility (user deletes drive)
1686	 * and deregister logical drives which have disappeared.
1687	 */
1688	for (i = 0; i <= h->highest_lun; i++) {
1689		int j;
1690		drv_found = 0;
1691		for (j = 0; j < num_luns; j++) {
1692			memcpy(&lunid, &ld_buff->LUN[j][0], 4);
1693			lunid = le32_to_cpu(lunid);
1694			if (h->drv[i].LunID == lunid) {
1695				drv_found = 1;
1696				break;
1697			}
1698		}
1699		if (!drv_found) {
1700			/* Deregister it from the OS, it's gone. */
1701			spin_lock_irqsave(CCISS_LOCK(h->ctlr), flags);
1702			h->drv[i].busy_configuring = 1;
1703			spin_unlock_irqrestore(CCISS_LOCK(h->ctlr), flags);
1704			return_code = deregister_disk(h->gendisk[i],
1705				&h->drv[i], 1);
1706			h->drv[i].busy_configuring = 0;
1707		}
1708	}
1709
1710	/* Compare controller drive array to driver's drive array.
1711	 * Check for updates in the drive information and any new drives
1712	 * on the controller due to ACU adding logical drives, or changing
1713	 * a logical drive's size, etc.  Reregister any new/changed drives
1714	 */
1715	for (i = 0; i < num_luns; i++) {
1716		int j;
1717
1718		drv_found = 0;
1719
1720		memcpy(&lunid, &ld_buff->LUN[i][0], 4);
1721		lunid = le32_to_cpu(lunid);
1722
1723		/* Find if the LUN is already in the drive array
1724		 * of the driver.  If so then update its info
1725		 * if not in use.  If it does not exist then find
1726		 * the first free index and add it.
1727		 */
1728		for (j = 0; j <= h->highest_lun; j++) {
1729			if (h->drv[j].raid_level != -1 &&
1730				h->drv[j].LunID == lunid) {
1731				drv_index = j;
1732				drv_found = 1;
1733				break;
1734			}
1735		}
1736
1737		/* check if the drive was found already in the array */
1738		if (!drv_found) {
1739			drv_index = cciss_add_gendisk(h, lunid, 0);
1740			if (drv_index == -1)
1741				goto freeret;
1742		}
1743		cciss_update_drive_info(ctlr, drv_index, first_time);
1744	}		/* end for */
1745
1746freeret:
1747	kfree(ld_buff);
1748	h->busy_configuring = 0;
1749	/* We return -1 here to tell the ACU that we have registered/updated
1750	 * all of the drives that we can and to keep it from calling us
1751	 * additional times.
1752	 */
1753	return -1;
1754mem_msg:
1755	printk(KERN_ERR "cciss: out of memory\n");
1756	h->busy_configuring = 0;
1757	goto freeret;
1758}
1759
1760/* This function will deregister the disk and it's queue from the
1761 * kernel.  It must be called with the controller lock held and the
1762 * drv structures busy_configuring flag set.  It's parameters are:
1763 *
1764 * disk = This is the disk to be deregistered
1765 * drv  = This is the drive_info_struct associated with the disk to be
1766 *        deregistered.  It contains information about the disk used
1767 *        by the driver.
1768 * clear_all = This flag determines whether or not the disk information
1769 *             is going to be completely cleared out and the highest_lun
1770 *             reset.  Sometimes we want to clear out information about
1771 *             the disk in preparation for re-adding it.  In this case
1772 *             the highest_lun should be left unchanged and the LunID
1773 *             should not be cleared.
1774*/
1775static int deregister_disk(struct gendisk *disk, drive_info_struct *drv,
1776			   int clear_all)
1777{
1778	int i;
1779	ctlr_info_t *h = get_host(disk);
1780
1781	if (!capable(CAP_SYS_RAWIO))
1782		return -EPERM;
1783
1784	/* make sure logical volume is NOT is use */
1785	if (clear_all || (h->gendisk[0] == disk)) {
1786		if (drv->usage_count > 1)
1787			return -EBUSY;
1788	} else if (drv->usage_count > 0)
1789		return -EBUSY;
1790
1791	/* invalidate the devices and deregister the disk.  If it is disk
1792	 * zero do not deregister it but just zero out it's values.  This
1793	 * allows us to delete disk zero but keep the controller registered.
1794	 */
1795	if (h->gendisk[0] != disk) {
1796		struct request_queue *q = disk->queue;
1797		if (disk->flags & GENHD_FL_UP)
1798			del_gendisk(disk);
1799		if (q) {
1800			blk_cleanup_queue(q);
1801			/* Set drv->queue to NULL so that we do not try
1802			 * to call blk_start_queue on this queue in the
1803			 * interrupt handler
1804			 */
1805			drv->queue = NULL;
1806		}
1807		/* If clear_all is set then we are deleting the logical
1808		 * drive, not just refreshing its info.  For drives
1809		 * other than disk 0 we will call put_disk.  We do not
1810		 * do this for disk 0 as we need it to be able to
1811		 * configure the controller.
1812		 */
1813		if (clear_all){
1814			/* This isn't pretty, but we need to find the
1815			 * disk in our array and NULL our the pointer.
1816			 * This is so that we will call alloc_disk if
1817			 * this index is used again later.
1818			 */
1819			for (i=0; i < CISS_MAX_LUN; i++){
1820				if (h->gendisk[i] == disk) {
1821					h->gendisk[i] = NULL;
1822					break;
1823				}
1824			}
1825			put_disk(disk);
1826		}
1827	} else {
1828		set_capacity(disk, 0);
1829	}
1830
1831	--h->num_luns;
1832	/* zero out the disk size info */
1833	drv->nr_blocks = 0;
1834	drv->block_size = 0;
1835	drv->heads = 0;
1836	drv->sectors = 0;
1837	drv->cylinders = 0;
1838	drv->raid_level = -1;	/* This can be used as a flag variable to
1839				 * indicate that this element of the drive
1840				 * array is free.
1841				 */
1842
1843	if (clear_all) {
1844		/* check to see if it was the last disk */
1845		if (drv == h->drv + h->highest_lun) {
1846			/* if so, find the new hightest lun */
1847			int i, newhighest = -1;
1848			for (i = 0; i <= h->highest_lun; i++) {
1849				/* if the disk has size > 0, it is available */
1850				if (h->drv[i].heads)
1851					newhighest = i;
1852			}
1853			h->highest_lun = newhighest;
1854		}
1855
1856		drv->LunID = 0;
1857	}
1858	return 0;
1859}
1860
1861static int fill_cmd(CommandList_struct *c, __u8 cmd, int ctlr, void *buff, size_t size, unsigned int use_unit_num,	/* 0: address the controller,
1862															   1: address logical volume log_unit,
1863															   2: periph device address is scsi3addr */
1864		    unsigned int log_unit, __u8 page_code,
1865		    unsigned char *scsi3addr, int cmd_type)
1866{
1867	ctlr_info_t *h = hba[ctlr];
1868	u64bit buff_dma_handle;
1869	int status = IO_OK;
1870
1871	c->cmd_type = CMD_IOCTL_PEND;
1872	c->Header.ReplyQueue = 0;
1873	if (buff != NULL) {
1874		c->Header.SGList = 1;
1875		c->Header.SGTotal = 1;
1876	} else {
1877		c->Header.SGList = 0;
1878		c->Header.SGTotal = 0;
1879	}
1880	c->Header.Tag.lower = c->busaddr;
1881
1882	c->Request.Type.Type = cmd_type;
1883	if (cmd_type == TYPE_CMD) {
1884		switch (cmd) {
1885		case CISS_INQUIRY:
1886			/* If the logical unit number is 0 then, this is going
1887			   to controller so It's a physical command
1888			   mode = 0 target = 0.  So we have nothing to write.
1889			   otherwise, if use_unit_num == 1,
1890			   mode = 1(volume set addressing) target = LUNID
1891			   otherwise, if use_unit_num == 2,
1892			   mode = 0(periph dev addr) target = scsi3addr */
1893			if (use_unit_num == 1) {
1894				c->Header.LUN.LogDev.VolId =
1895				    h->drv[log_unit].LunID;
1896				c->Header.LUN.LogDev.Mode = 1;
1897			} else if (use_unit_num == 2) {
1898				memcpy(c->Header.LUN.LunAddrBytes, scsi3addr,
1899				       8);
1900				c->Header.LUN.LogDev.Mode = 0;
1901			}
1902			/* are we trying to read a vital product page */
1903			if (page_code != 0) {
1904				c->Request.CDB[1] = 0x01;
1905				c->Request.CDB[2] = page_code;
1906			}
1907			c->Request.CDBLen = 6;
1908			c->Request.Type.Attribute = ATTR_SIMPLE;
1909			c->Request.Type.Direction = XFER_READ;
1910			c->Request.Timeout = 0;
1911			c->Request.CDB[0] = CISS_INQUIRY;
1912			c->Request.CDB[4] = size & 0xFF;
1913			break;
1914		case CISS_REPORT_LOG:
1915		case CISS_REPORT_PHYS:
1916			/* Talking to controller so It's a physical command
1917			   mode = 00 target = 0.  Nothing to write.
1918			 */
1919			c->Request.CDBLen = 12;
1920			c->Request.Type.Attribute = ATTR_SIMPLE;
1921			c->Request.Type.Direction = XFER_READ;
1922			c->Request.Timeout = 0;
1923			c->Request.CDB[0] = cmd;
1924			c->Request.CDB[6] = (size >> 24) & 0xFF;	//MSB
1925			c->Request.CDB[7] = (size >> 16) & 0xFF;
1926			c->Request.CDB[8] = (size >> 8) & 0xFF;
1927			c->Request.CDB[9] = size & 0xFF;
1928			break;
1929
1930		case CCISS_READ_CAPACITY:
1931			c->Header.LUN.LogDev.VolId = h->drv[log_unit].LunID;
1932			c->Header.LUN.LogDev.Mode = 1;
1933			c->Request.CDBLen = 10;
1934			c->Request.Type.Attribute = ATTR_SIMPLE;
1935			c->Request.Type.Direction = XFER_READ;
1936			c->Request.Timeout = 0;
1937			c->Request.CDB[0] = cmd;
1938			break;
1939		case CCISS_READ_CAPACITY_16:
1940			c->Header.LUN.LogDev.VolId = h->drv[log_unit].LunID;
1941			c->Header.LUN.LogDev.Mode = 1;
1942			c->Request.CDBLen = 16;
1943			c->Request.Type.Attribute = ATTR_SIMPLE;
1944			c->Request.Type.Direction = XFER_READ;
1945			c->Request.Timeout = 0;
1946			c->Request.CDB[0] = cmd;
1947			c->Request.CDB[1] = 0x10;
1948			c->Request.CDB[10] = (size >> 24) & 0xFF;
1949			c->Request.CDB[11] = (size >> 16) & 0xFF;
1950			c->Request.CDB[12] = (size >> 8) & 0xFF;
1951			c->Request.CDB[13] = size & 0xFF;
1952			c->Request.Timeout = 0;
1953			c->Request.CDB[0] = cmd;
1954			break;
1955		case CCISS_CACHE_FLUSH:
1956			c->Request.CDBLen = 12;
1957			c->Request.Type.Attribute = ATTR_SIMPLE;
1958			c->Request.Type.Direction = XFER_WRITE;
1959			c->Request.Timeout = 0;
1960			c->Request.CDB[0] = BMIC_WRITE;
1961			c->Request.CDB[6] = BMIC_CACHE_FLUSH;
1962			break;
1963		default:
1964			printk(KERN_WARNING
1965			       "cciss%d:  Unknown Command 0x%c\n", ctlr, cmd);
1966			return IO_ERROR;
1967		}
1968	} else if (cmd_type == TYPE_MSG) {
1969		switch (cmd) {
1970		case 0:	/* ABORT message */
1971			c->Request.CDBLen = 12;
1972			c->Request.Type.Attribute = ATTR_SIMPLE;
1973			c->Request.Type.Direction = XFER_WRITE;
1974			c->Request.Timeout = 0;
1975			c->Request.CDB[0] = cmd;	/* abort */
1976			c->Request.CDB[1] = 0;	/* abort a command */
1977			/* buff contains the tag of the command to abort */
1978			memcpy(&c->Request.CDB[4], buff, 8);
1979			break;
1980		case 1:	/* RESET message */
1981			c->Request.CDBLen = 12;
1982			c->Request.Type.Attribute = ATTR_SIMPLE;
1983			c->Request.Type.Direction = XFER_WRITE;
1984			c->Request.Timeout = 0;
1985			memset(&c->Request.CDB[0], 0, sizeof(c->Request.CDB));
1986			c->Request.CDB[0] = cmd;	/* reset */
1987			c->Request.CDB[1] = 0x04;	/* reset a LUN */
1988			break;
1989		case 3:	/* No-Op message */
1990			c->Request.CDBLen = 1;
1991			c->Request.Type.Attribute = ATTR_SIMPLE;
1992			c->Request.Type.Direction = XFER_WRITE;
1993			c->Request.Timeout = 0;
1994			c->Request.CDB[0] = cmd;
1995			break;
1996		default:
1997			printk(KERN_WARNING
1998			       "cciss%d: unknown message type %d\n", ctlr, cmd);
1999			return IO_ERROR;
2000		}
2001	} else {
2002		printk(KERN_WARNING
2003		       "cciss%d: unknown command type %d\n", ctlr, cmd_type);
2004		return IO_ERROR;
2005	}
2006	/* Fill in the scatter gather information */
2007	if (size > 0) {
2008		buff_dma_handle.val = (__u64) pci_map_single(h->pdev,
2009							     buff, size,
2010							     PCI_DMA_BIDIRECTIONAL);
2011		c->SG[0].Addr.lower = buff_dma_handle.val32.lower;
2012		c->SG[0].Addr.upper = buff_dma_handle.val32.upper;
2013		c->SG[0].Len = size;
2014		c->SG[0].Ext = 0;	/* we are not chaining */
2015	}
2016	return status;
2017}
2018
2019static int sendcmd_withirq(__u8 cmd,
2020			   int ctlr,
2021			   void *buff,
2022			   size_t size,
2023			   unsigned int use_unit_num,
2024			   unsigned int log_unit, __u8 page_code, int cmd_type)
2025{
2026	ctlr_info_t *h = hba[ctlr];
2027	CommandList_struct *c;
2028	u64bit buff_dma_handle;
2029	unsigned long flags;
2030	int return_status;
2031	DECLARE_COMPLETION_ONSTACK(wait);
2032
2033	if ((c = cmd_alloc(h, 0)) == NULL)
2034		return -ENOMEM;
2035	return_status = fill_cmd(c, cmd, ctlr, buff, size, use_unit_num,
2036				 log_unit, page_code, NULL, cmd_type);
2037	if (return_status != IO_OK) {
2038		cmd_free(h, c, 0);
2039		return return_status;
2040	}
2041      resend_cmd2:
2042	c->waiting = &wait;
2043
2044	/* Put the request on the tail of the queue and send it */
2045	spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
2046	addQ(&h->reqQ, c);
2047	h->Qdepth++;
2048	start_io(h);
2049	spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
2050
2051	wait_for_completion(&wait);
2052
2053	if (c->err_info->CommandStatus != 0) {	/* an error has occurred */
2054		switch (c->err_info->CommandStatus) {
2055		case CMD_TARGET_STATUS:
2056			printk(KERN_WARNING "cciss: cmd %p has "
2057			       " completed with errors\n", c);
2058			if (c->err_info->ScsiStatus) {
2059				printk(KERN_WARNING "cciss: cmd %p "
2060				       "has SCSI Status = %x\n",
2061				       c, c->err_info->ScsiStatus);
2062			}
2063
2064			break;
2065		case CMD_DATA_UNDERRUN:
2066		case CMD_DATA_OVERRUN:
2067			/* expected for inquire and report lun commands */
2068			break;
2069		case CMD_INVALID:
2070			printk(KERN_WARNING "cciss: Cmd %p is "
2071			       "reported invalid\n", c);
2072			return_status = IO_ERROR;
2073			break;
2074		case CMD_PROTOCOL_ERR:
2075			printk(KERN_WARNING "cciss: cmd %p has "
2076			       "protocol error \n", c);
2077			return_status = IO_ERROR;
2078			break;
2079		case CMD_HARDWARE_ERR:
2080			printk(KERN_WARNING "cciss: cmd %p had "
2081			       " hardware error\n", c);
2082			return_status = IO_ERROR;
2083			break;
2084		case CMD_CONNECTION_LOST:
2085			printk(KERN_WARNING "cciss: cmd %p had "
2086			       "connection lost\n", c);
2087			return_status = IO_ERROR;
2088			break;
2089		case CMD_ABORTED:
2090			printk(KERN_WARNING "cciss: cmd %p was "
2091			       "aborted\n", c);
2092			return_status = IO_ERROR;
2093			break;
2094		case CMD_ABORT_FAILED:
2095			printk(KERN_WARNING "cciss: cmd %p reports "
2096			       "abort failed\n", c);
2097			return_status = IO_ERROR;
2098			break;
2099		case CMD_UNSOLICITED_ABORT:
2100			printk(KERN_WARNING
2101			       "cciss%d: unsolicited abort %p\n", ctlr, c);
2102			if (c->retry_count < MAX_CMD_RETRIES) {
2103				printk(KERN_WARNING
2104				       "cciss%d: retrying %p\n", ctlr, c);
2105				c->retry_count++;
2106				/* erase the old error information */
2107				memset(c->err_info, 0,
2108				       sizeof(ErrorInfo_struct));
2109				return_status = IO_OK;
2110				INIT_COMPLETION(wait);
2111				goto resend_cmd2;
2112			}
2113			return_status = IO_ERROR;
2114			break;
2115		default:
2116			printk(KERN_WARNING "cciss: cmd %p returned "
2117			       "unknown status %x\n", c,
2118			       c->err_info->CommandStatus);
2119			return_status = IO_ERROR;
2120		}
2121	}
2122	/* unlock the buffers from DMA */
2123	buff_dma_handle.val32.lower = c->SG[0].Addr.lower;
2124	buff_dma_handle.val32.upper = c->SG[0].Addr.upper;
2125	pci_unmap_single(h->pdev, (dma_addr_t) buff_dma_handle.val,
2126			 c->SG[0].Len, PCI_DMA_BIDIRECTIONAL);
2127	cmd_free(h, c, 0);
2128	return return_status;
2129}
2130
2131static void cciss_geometry_inquiry(int ctlr, int logvol,
2132				   int withirq, sector_t total_size,
2133				   unsigned int block_size,
2134				   InquiryData_struct *inq_buff,
2135				   drive_info_struct *drv)
2136{
2137	int return_code;
2138	unsigned long t;
2139
2140	memset(inq_buff, 0, sizeof(InquiryData_struct));
2141	if (withirq)
2142		return_code = sendcmd_withirq(CISS_INQUIRY, ctlr,
2143					      inq_buff, sizeof(*inq_buff), 1,
2144					      logvol, 0xC1, TYPE_CMD);
2145	else
2146		return_code = sendcmd(CISS_INQUIRY, ctlr, inq_buff,
2147				      sizeof(*inq_buff), 1, logvol, 0xC1, NULL,
2148				      TYPE_CMD);
2149	if (return_code == IO_OK) {
2150		if (inq_buff->data_byte[8] == 0xFF) {
2151			printk(KERN_WARNING
2152			       "cciss: reading geometry failed, volume "
2153			       "does not support reading geometry\n");
2154			drv->heads = 255;
2155			drv->sectors = 32;	// Sectors per track
2156			drv->cylinders = total_size + 1;
2157			drv->raid_level = RAID_UNKNOWN;
2158		} else {
2159			drv->heads = inq_buff->data_byte[6];
2160			drv->sectors = inq_buff->data_byte[7];
2161			drv->cylinders = (inq_buff->data_byte[4] & 0xff) << 8;
2162			drv->cylinders += inq_buff->data_byte[5];
2163			drv->raid_level = inq_buff->data_byte[8];
2164		}
2165		drv->block_size = block_size;
2166		drv->nr_blocks = total_size + 1;
2167		t = drv->heads * drv->sectors;
2168		if (t > 1) {
2169			sector_t real_size = total_size + 1;
2170			unsigned long rem = sector_div(real_size, t);
2171			if (rem)
2172				real_size++;
2173			drv->cylinders = real_size;
2174		}
2175	} else {		/* Get geometry failed */
2176		printk(KERN_WARNING "cciss: reading geometry failed\n");
2177	}
2178	printk(KERN_INFO "      heads=%d, sectors=%d, cylinders=%d\n\n",
2179	       drv->heads, drv->sectors, drv->cylinders);
2180}
2181
2182static void
2183cciss_read_capacity(int ctlr, int logvol, int withirq, sector_t *total_size,
2184		    unsigned int *block_size)
2185{
2186	ReadCapdata_struct *buf;
2187	int return_code;
2188
2189	buf = kzalloc(sizeof(ReadCapdata_struct), GFP_KERNEL);
2190	if (!buf) {
2191		printk(KERN_WARNING "cciss: out of memory\n");
2192		return;
2193	}
2194
2195	if (withirq)
2196		return_code = sendcmd_withirq(CCISS_READ_CAPACITY,
2197				ctlr, buf, sizeof(ReadCapdata_struct),
2198					1, logvol, 0, TYPE_CMD);
2199	else
2200		return_code = sendcmd(CCISS_READ_CAPACITY,
2201				ctlr, buf, sizeof(ReadCapdata_struct),
2202					1, logvol, 0, NULL, TYPE_CMD);
2203	if (return_code == IO_OK) {
2204		*total_size = be32_to_cpu(*(__be32 *) buf->total_size);
2205		*block_size = be32_to_cpu(*(__be32 *) buf->block_size);
2206	} else {		/* read capacity command failed */
2207		printk(KERN_WARNING "cciss: read capacity failed\n");
2208		*total_size = 0;
2209		*block_size = BLOCK_SIZE;
2210	}
2211	if (*total_size != 0)
2212		printk(KERN_INFO "      blocks= %llu block_size= %d\n",
2213		(unsigned long long)*total_size+1, *block_size);
2214	kfree(buf);
2215}
2216
2217static void
2218cciss_read_capacity_16(int ctlr, int logvol, int withirq, sector_t *total_size, 				unsigned int *block_size)
2219{
2220	ReadCapdata_struct_16 *buf;
2221	int return_code;
2222
2223	buf = kzalloc(sizeof(ReadCapdata_struct_16), GFP_KERNEL);
2224	if (!buf) {
2225		printk(KERN_WARNING "cciss: out of memory\n");
2226		return;
2227	}
2228
2229	if (withirq) {
2230		return_code = sendcmd_withirq(CCISS_READ_CAPACITY_16,
2231			ctlr, buf, sizeof(ReadCapdata_struct_16),
2232				1, logvol, 0, TYPE_CMD);
2233	}
2234	else {
2235		return_code = sendcmd(CCISS_READ_CAPACITY_16,
2236			ctlr, buf, sizeof(ReadCapdata_struct_16),
2237				1, logvol, 0, NULL, TYPE_CMD);
2238	}
2239	if (return_code == IO_OK) {
2240		*total_size = be64_to_cpu(*(__be64 *) buf->total_size);
2241		*block_size = be32_to_cpu(*(__be32 *) buf->block_size);
2242	} else {		/* read capacity command failed */
2243		printk(KERN_WARNING "cciss: read capacity failed\n");
2244		*total_size = 0;
2245		*block_size = BLOCK_SIZE;
2246	}
2247	printk(KERN_INFO "      blocks= %llu block_size= %d\n",
2248	       (unsigned long long)*total_size+1, *block_size);
2249	kfree(buf);
2250}
2251
2252static int cciss_revalidate(struct gendisk *disk)
2253{
2254	ctlr_info_t *h = get_host(disk);
2255	drive_info_struct *drv = get_drv(disk);
2256	int logvol;
2257	int FOUND = 0;
2258	unsigned int block_size;
2259	sector_t total_size;
2260	InquiryData_struct *inq_buff = NULL;
2261
2262	for (logvol = 0; logvol < CISS_MAX_LUN; logvol++) {
2263		if (h->drv[logvol].LunID == drv->LunID) {
2264			FOUND = 1;
2265			break;
2266		}
2267	}
2268
2269	if (!FOUND)
2270		return 1;
2271
2272	inq_buff = kmalloc(sizeof(InquiryData_struct), GFP_KERNEL);
2273	if (inq_buff == NULL) {
2274		printk(KERN_WARNING "cciss: out of memory\n");
2275		return 1;
2276	}
2277	if (h->cciss_read == CCISS_READ_10) {
2278		cciss_read_capacity(h->ctlr, logvol, 1,
2279					&total_size, &block_size);
2280	} else {
2281		cciss_read_capacity_16(h->ctlr, logvol, 1,
2282					&total_size, &block_size);
2283	}
2284	cciss_geometry_inquiry(h->ctlr, logvol, 1, total_size, block_size,
2285			       inq_buff, drv);
2286
2287	blk_queue_hardsect_size(drv->queue, drv->block_size);
2288	set_capacity(disk, drv->nr_blocks);
2289
2290	kfree(inq_buff);
2291	return 0;
2292}
2293
2294/*
2295 *   Wait polling for a command to complete.
2296 *   The memory mapped FIFO is polled for the completion.
2297 *   Used only at init time, interrupts from the HBA are disabled.
2298 */
2299static unsigned long pollcomplete(int ctlr)
2300{
2301	unsigned long done;
2302	int i;
2303
2304	/* Wait (up to 20 seconds) for a command to complete */
2305
2306	for (i = 20 * HZ; i > 0; i--) {
2307		done = hba[ctlr]->access.command_completed(hba[ctlr]);
2308		if (done == FIFO_EMPTY)
2309			schedule_timeout_uninterruptible(1);
2310		else
2311			return done;
2312	}
2313	/* Invalid address to tell caller we ran out of time */
2314	return 1;
2315}
2316
2317static int add_sendcmd_reject(__u8 cmd, int ctlr, unsigned long complete)
2318{
2319	/* We get in here if sendcmd() is polling for completions
2320	   and gets some command back that it wasn't expecting --
2321	   something other than that which it just sent down.
2322	   Ordinarily, that shouldn't happen, but it can happen when
2323	   the scsi tape stuff gets into error handling mode, and
2324	   starts using sendcmd() to try to abort commands and
2325	   reset tape drives.  In that case, sendcmd may pick up
2326	   completions of commands that were sent to logical drives
2327	   through the block i/o system, or cciss ioctls completing, etc.
2328	   In that case, we need to save those completions for later
2329	   processing by the interrupt handler.
2330	 */
2331
2332#ifdef CONFIG_CISS_SCSI_TAPE
2333	struct sendcmd_reject_list *srl = &hba[ctlr]->scsi_rejects;
2334
2335	/* If it's not the scsi tape stuff doing error handling, (abort */
2336	/* or reset) then we don't expect anything weird. */
2337	if (cmd != CCISS_RESET_MSG && cmd != CCISS_ABORT_MSG) {
2338#endif
2339		printk(KERN_WARNING "cciss cciss%d: SendCmd "
2340		       "Invalid command list address returned! (%lx)\n",
2341		       ctlr, complete);
2342		/* not much we can do. */
2343#ifdef CONFIG_CISS_SCSI_TAPE
2344		return 1;
2345	}
2346
2347	/* We've sent down an abort or reset, but something else
2348	   has completed */
2349	if (srl->ncompletions >= (hba[ctlr]->nr_cmds + 2)) {
2350		/* Uh oh.  No room to save it for later... */
2351		printk(KERN_WARNING "cciss%d: Sendcmd: Invalid command addr, "
2352		       "reject list overflow, command lost!\n", ctlr);
2353		return 1;
2354	}
2355	/* Save it for later */
2356	srl->complete[srl->ncompletions] = complete;
2357	srl->ncompletions++;
2358#endif
2359	return 0;
2360}
2361
2362/*
2363 * Send a command to the controller, and wait for it to complete.
2364 * Only used at init time.
2365 */
2366static int sendcmd(__u8 cmd, int ctlr, void *buff, size_t size, unsigned int use_unit_num,	/* 0: address the controller,
2367												   1: address logical volume log_unit,
2368												   2: periph device address is scsi3addr */
2369		   unsigned int log_unit,
2370		   __u8 page_code, unsigned char *scsi3addr, int cmd_type)
2371{
2372	CommandList_struct *c;
2373	int i;
2374	unsigned long complete;
2375	ctlr_info_t *info_p = hba[ctlr];
2376	u64bit buff_dma_handle;
2377	int status, done = 0;
2378
2379	if ((c = cmd_alloc(info_p, 1)) == NULL) {
2380		printk(KERN_WARNING "cciss: unable to get memory");
2381		return IO_ERROR;
2382	}
2383	status = fill_cmd(c, cmd, ctlr, buff, size, use_unit_num,
2384			  log_unit, page_code, scsi3addr, cmd_type);
2385	if (status != IO_OK) {
2386		cmd_free(info_p, c, 1);
2387		return status;
2388	}
2389      resend_cmd1:
2390	/*
2391	 * Disable interrupt
2392	 */
2393#ifdef CCISS_DEBUG
2394	printk(KERN_DEBUG "cciss: turning intr off\n");
2395#endif				/* CCISS_DEBUG */
2396	info_p->access.set_intr_mask(info_p, CCISS_INTR_OFF);
2397
2398	/* Make sure there is room in the command FIFO */
2399	/* Actually it should be completely empty at this time */
2400	/* unless we are in here doing error handling for the scsi */
2401	/* tape side of the driver. */
2402	for (i = 200000; i > 0; i--) {
2403		/* if fifo isn't full go */
2404		if (!(info_p->access.fifo_full(info_p))) {
2405
2406			break;
2407		}
2408		udelay(10);
2409		printk(KERN_WARNING "cciss cciss%d: SendCmd FIFO full,"
2410		       " waiting!\n", ctlr);
2411	}
2412	/*
2413	 * Send the cmd
2414	 */
2415	info_p->access.submit_command(info_p, c);
2416	done = 0;
2417	do {
2418		complete = pollcomplete(ctlr);
2419
2420#ifdef CCISS_DEBUG
2421		printk(KERN_DEBUG "cciss: command completed\n");
2422#endif				/* CCISS_DEBUG */
2423
2424		if (complete == 1) {
2425			printk(KERN_WARNING
2426			       "cciss cciss%d: SendCmd Timeout out, "
2427			       "No command list address returned!\n", ctlr);
2428			status = IO_ERROR;
2429			done = 1;
2430			break;
2431		}
2432
2433		/* This will need to change for direct lookup completions */
2434		if ((complete & CISS_ERROR_BIT)
2435		    && (complete & ~CISS_ERROR_BIT) == c->busaddr) {
2436			/* if data overrun or underun on Report command
2437			   ignore it
2438			 */
2439			if (((c->Request.CDB[0] == CISS_REPORT_LOG) ||
2440			     (c->Request.CDB[0] == CISS_REPORT_PHYS) ||
2441			     (c->Request.CDB[0] == CISS_INQUIRY)) &&
2442			    ((c->err_info->CommandStatus ==
2443			      CMD_DATA_OVERRUN) ||
2444			     (c->err_info->CommandStatus == CMD_DATA_UNDERRUN)
2445			    )) {
2446				complete = c->busaddr;
2447			} else {
2448				if (c->err_info->CommandStatus ==
2449				    CMD_UNSOLICITED_ABORT) {
2450					printk(KERN_WARNING "cciss%d: "
2451					       "unsolicited abort %p\n",
2452					       ctlr, c);
2453					if (c->retry_count < MAX_CMD_RETRIES) {
2454						printk(KERN_WARNING
2455						       "cciss%d: retrying %p\n",
2456						       ctlr, c);
2457						c->retry_count++;
2458						/* erase the old error */
2459						/* information */
2460						memset(c->err_info, 0,
2461						       sizeof
2462						       (ErrorInfo_struct));
2463						goto resend_cmd1;
2464					} else {
2465						printk(KERN_WARNING
2466						       "cciss%d: retried %p too "
2467						       "many times\n", ctlr, c);
2468						status = IO_ERROR;
2469						goto cleanup1;
2470					}
2471				} else if (c->err_info->CommandStatus ==
2472					   CMD_UNABORTABLE) {
2473					printk(KERN_WARNING
2474					       "cciss%d: command could not be aborted.\n",
2475					       ctlr);
2476					status = IO_ERROR;
2477					goto cleanup1;
2478				}
2479				printk(KERN_WARNING "ciss ciss%d: sendcmd"
2480				       " Error %x \n", ctlr,
2481				       c->err_info->CommandStatus);
2482				printk(KERN_WARNING "ciss ciss%d: sendcmd"
2483				       " offensive info\n"
2484				       "  size %x\n   num %x   value %x\n",
2485				       ctlr,
2486				       c->err_info->MoreErrInfo.Invalid_Cmd.
2487				       offense_size,
2488				       c->err_info->MoreErrInfo.Invalid_Cmd.
2489				       offense_num,
2490				       c->err_info->MoreErrInfo.Invalid_Cmd.
2491				       offense_value);
2492				status = IO_ERROR;
2493				goto cleanup1;
2494			}
2495		}
2496		/* This will need changing for direct lookup completions */
2497		if (complete != c->busaddr) {
2498			if (add_sendcmd_reject(cmd, ctlr, complete) != 0) {
2499				BUG();	/* we are pretty much hosed if we get here. */
2500			}
2501			continue;
2502		} else
2503			done = 1;
2504	} while (!done);
2505
2506      cleanup1:
2507	/* unlock the data buffer from DMA */
2508	buff_dma_handle.val32.lower = c->SG[0].Addr.lower;
2509	buff_dma_handle.val32.upper = c->SG[0].Addr.upper;
2510	pci_unmap_single(info_p->pdev, (dma_addr_t) buff_dma_handle.val,
2511			 c->SG[0].Len, PCI_DMA_BIDIRECTIONAL);
2512#ifdef CONFIG_CISS_SCSI_TAPE
2513	/* if we saved some commands for later, process them now. */
2514	if (info_p->scsi_rejects.ncompletions > 0)
2515		do_cciss_intr(0, info_p);
2516#endif
2517	cmd_free(info_p, c, 1);
2518	return status;
2519}
2520
2521/*
2522 * Map (physical) PCI mem into (virtual) kernel space
2523 */
2524static void __iomem *remap_pci_mem(ulong base, ulong size)
2525{
2526	ulong page_base = ((ulong) base) & PAGE_MASK;
2527	ulong page_offs = ((ulong) base) - page_base;
2528	void __iomem *page_remapped = ioremap(page_base, page_offs + size);
2529
2530	return page_remapped ? (page_remapped + page_offs) : NULL;
2531}
2532
2533/*
2534 * Takes jobs of the Q and sends them to the hardware, then puts it on
2535 * the Q to wait for completion.
2536 */
2537static void start_io(ctlr_info_t *h)
2538{
2539	CommandList_struct *c;
2540
2541	while ((c = h->reqQ) != NULL) {
2542		/* can't do anything if fifo is full */
2543		if ((h->access.fifo_full(h))) {
2544			printk(KERN_WARNING "cciss: fifo full\n");
2545			break;
2546		}
2547
2548		/* Get the first entry from the Request Q */
2549		removeQ(&(h->reqQ), c);
2550		h->Qdepth--;
2551
2552		/* Tell the controller execute command */
2553		h->access.submit_command(h, c);
2554
2555		/* Put job onto the completed Q */
2556		addQ(&(h->cmpQ), c);
2557	}
2558}
2559
2560/* Assumes that CCISS_LOCK(h->ctlr) is held. */
2561/* Zeros out the error record and then resends the command back */
2562/* to the controller */
2563static inline void resend_cciss_cmd(ctlr_info_t *h, CommandList_struct *c)
2564{
2565	/* erase the old error information */
2566	memset(c->err_info, 0, sizeof(ErrorInfo_struct));
2567
2568	/* add it to software queue and then send it to the controller */
2569	addQ(&(h->reqQ), c);
2570	h->Qdepth++;
2571	if (h->Qdepth > h->maxQsinceinit)
2572		h->maxQsinceinit = h->Qdepth;
2573
2574	start_io(h);
2575}
2576
2577static inline unsigned int make_status_bytes(unsigned int scsi_status_byte,
2578	unsigned int msg_byte, unsigned int host_byte,
2579	unsigned int driver_byte)
2580{
2581	/* inverse of macros in scsi.h */
2582	return (scsi_status_byte & 0xff) |
2583		((msg_byte & 0xff) << 8) |
2584		((host_byte & 0xff) << 16) |
2585		((driver_byte & 0xff) << 24);
2586}
2587
2588static inline int evaluate_target_status(CommandList_struct *cmd)
2589{
2590	unsigned char sense_key;
2591	unsigned char status_byte, msg_byte, host_byte, driver_byte;
2592	int error_value;
2593
2594	/* If we get in here, it means we got "target status", that is, scsi status */
2595	status_byte = cmd->err_info->ScsiStatus;
2596	driver_byte = DRIVER_OK;
2597	msg_byte = cmd->err_info->CommandStatus; /* correct?  seems too device specific */
2598
2599	if (blk_pc_request(cmd->rq))
2600		host_byte = DID_PASSTHROUGH;
2601	else
2602		host_byte = DID_OK;
2603
2604	error_value = make_status_bytes(status_byte, msg_byte,
2605		host_byte, driver_byte);
2606
2607	if (cmd->err_info->ScsiStatus != SAM_STAT_CHECK_CONDITION) {
2608		if (!blk_pc_request(cmd->rq))
2609			printk(KERN_WARNING "cciss: cmd %p "
2610			       "has SCSI Status 0x%x\n",
2611			       cmd, cmd->err_info->ScsiStatus);
2612		return error_value;
2613	}
2614
2615	/* check the sense key */
2616	sense_key = 0xf & cmd->err_info->SenseInfo[2];
2617	/* no status or recovered error */
2618	if (((sense_key == 0x0) || (sense_key == 0x1)) && !blk_pc_request(cmd->rq))
2619		error_value = 0;
2620
2621	if (!blk_pc_request(cmd->rq)) { /* Not SG_IO or similar? */
2622		if (error_value != 0)
2623			printk(KERN_WARNING "cciss: cmd %p has CHECK CONDITION"
2624			       " sense key = 0x%x\n", cmd, sense_key);
2625		return error_value;
2626	}
2627
2628	/* SG_IO or similar, copy sense data back */
2629	if (cmd->rq->sense) {
2630		if (cmd->rq->sense_len > cmd->err_info->SenseLen)
2631			cmd->rq->sense_len = cmd->err_info->SenseLen;
2632		memcpy(cmd->rq->sense, cmd->err_info->SenseInfo,
2633			cmd->rq->sense_len);
2634	} else
2635		cmd->rq->sense_len = 0;
2636
2637	return error_value;
2638}
2639
2640/* checks the status of the job and calls complete buffers to mark all
2641 * buffers for the completed job. Note that this function does not need
2642 * to hold the hba/queue lock.
2643 */
2644static inline void complete_command(ctlr_info_t *h, CommandList_struct *cmd,
2645				    int timeout)
2646{
2647	int retry_cmd = 0;
2648	struct request *rq = cmd->rq;
2649
2650	rq->errors = 0;
2651
2652	if (timeout)
2653		rq->errors = make_status_bytes(0, 0, 0, DRIVER_TIMEOUT);
2654
2655	if (cmd->err_info->CommandStatus == 0)	/* no error has occurred */
2656		goto after_error_processing;
2657
2658	switch (cmd->err_info->CommandStatus) {
2659	case CMD_TARGET_STATUS:
2660		rq->errors = evaluate_target_status(cmd);
2661		break;
2662	case CMD_DATA_UNDERRUN:
2663		if (blk_fs_request(cmd->rq)) {
2664			printk(KERN_WARNING "cciss: cmd %p has"
2665			       " completed with data underrun "
2666			       "reported\n", cmd);
2667			cmd->rq->data_len = cmd->err_info->ResidualCnt;
2668		}
2669		break;
2670	case CMD_DATA_OVERRUN:
2671		if (blk_fs_request(cmd->rq))
2672			printk(KERN_WARNING "cciss: cmd %p has"
2673			       " completed with data overrun "
2674			       "reported\n", cmd);
2675		break;
2676	case CMD_INVALID:
2677		printk(KERN_WARNING "cciss: cmd %p is "
2678		       "reported invalid\n", cmd);
2679		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2680			cmd->err_info->CommandStatus, DRIVER_OK,
2681			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2682		break;
2683	case CMD_PROTOCOL_ERR:
2684		printk(KERN_WARNING "cciss: cmd %p has "
2685		       "protocol error \n", cmd);
2686		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2687			cmd->err_info->CommandStatus, DRIVER_OK,
2688			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2689		break;
2690	case CMD_HARDWARE_ERR:
2691		printk(KERN_WARNING "cciss: cmd %p had "
2692		       " hardware error\n", cmd);
2693		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2694			cmd->err_info->CommandStatus, DRIVER_OK,
2695			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2696		break;
2697	case CMD_CONNECTION_LOST:
2698		printk(KERN_WARNING "cciss: cmd %p had "
2699		       "connection lost\n", cmd);
2700		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2701			cmd->err_info->CommandStatus, DRIVER_OK,
2702			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2703		break;
2704	case CMD_ABORTED:
2705		printk(KERN_WARNING "cciss: cmd %p was "
2706		       "aborted\n", cmd);
2707		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2708			cmd->err_info->CommandStatus, DRIVER_OK,
2709			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ABORT);
2710		break;
2711	case CMD_ABORT_FAILED:
2712		printk(KERN_WARNING "cciss: cmd %p reports "
2713		       "abort failed\n", cmd);
2714		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2715			cmd->err_info->CommandStatus, DRIVER_OK,
2716			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2717		break;
2718	case CMD_UNSOLICITED_ABORT:
2719		printk(KERN_WARNING "cciss%d: unsolicited "
2720		       "abort %p\n", h->ctlr, cmd);
2721		if (cmd->retry_count < MAX_CMD_RETRIES) {
2722			retry_cmd = 1;
2723			printk(KERN_WARNING
2724			       "cciss%d: retrying %p\n", h->ctlr, cmd);
2725			cmd->retry_count++;
2726		} else
2727			printk(KERN_WARNING
2728			       "cciss%d: %p retried too "
2729			       "many times\n", h->ctlr, cmd);
2730		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2731			cmd->err_info->CommandStatus, DRIVER_OK,
2732			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ABORT);
2733		break;
2734	case CMD_TIMEOUT:
2735		printk(KERN_WARNING "cciss: cmd %p timedout\n", cmd);
2736		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2737			cmd->err_info->CommandStatus, DRIVER_OK,
2738			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2739		break;
2740	default:
2741		printk(KERN_WARNING "cciss: cmd %p returned "
2742		       "unknown status %x\n", cmd,
2743		       cmd->err_info->CommandStatus);
2744		rq->errors = make_status_bytes(SAM_STAT_GOOD,
2745			cmd->err_info->CommandStatus, DRIVER_OK,
2746			blk_pc_request(cmd->rq) ? DID_PASSTHROUGH : DID_ERROR);
2747	}
2748
2749after_error_processing:
2750
2751	/* We need to return this command */
2752	if (retry_cmd) {
2753		resend_cciss_cmd(h, cmd);
2754		return;
2755	}
2756	cmd->rq->completion_data = cmd;
2757	blk_complete_request(cmd->rq);
2758}
2759
2760/*
2761 * Get a request and submit it to the controller.
2762 */
2763static void do_cciss_request(struct request_queue *q)
2764{
2765	ctlr_info_t *h = q->queuedata;
2766	CommandList_struct *c;
2767	sector_t start_blk;
2768	int seg;
2769	struct request *creq;
2770	u64bit temp64;
2771	struct scatterlist tmp_sg[MAXSGENTRIES];
2772	drive_info_struct *drv;
2773	int i, dir;
2774
2775	/* We call start_io here in case there is a command waiting on the
2776	 * queue that has not been sent.
2777	 */
2778	if (blk_queue_plugged(q))
2779		goto startio;
2780
2781      queue:
2782	creq = elv_next_request(q);
2783	if (!creq)
2784		goto startio;
2785
2786	BUG_ON(creq->nr_phys_segments > MAXSGENTRIES);
2787
2788	if ((c = cmd_alloc(h, 1)) == NULL)
2789		goto full;
2790
2791	blkdev_dequeue_request(creq);
2792
2793	spin_unlock_irq(q->queue_lock);
2794
2795	c->cmd_type = CMD_RWREQ;
2796	c->rq = creq;
2797
2798	/* fill in the request */
2799	drv = creq->rq_disk->private_data;
2800	c->Header.ReplyQueue = 0;	// unused in simple mode
2801	/* got command from pool, so use the command block index instead */
2802	/* for direct lookups. */
2803	/* The first 2 bits are reserved for controller error reporting. */
2804	c->Header.Tag.lower = (c->cmdindex << 3);
2805	c->Header.Tag.lower |= 0x04;	/* flag for direct lookup. */
2806	c->Header.LUN.LogDev.VolId = drv->LunID;
2807	c->Header.LUN.LogDev.Mode = 1;
2808	c->Request.CDBLen = 10;	// 12 byte commands not in FW yet;
2809	c->Request.Type.Type = TYPE_CMD;	// It is a command.
2810	c->Request.Type.Attribute = ATTR_SIMPLE;
2811	c->Request.Type.Direction =
2812	    (rq_data_dir(creq) == READ) ? XFER_READ : XFER_WRITE;
2813	c->Request.Timeout = 0;	// Don't time out
2814	c->Request.CDB[0] =
2815	    (rq_data_dir(creq) == READ) ? h->cciss_read : h->cciss_write;
2816	start_blk = creq->sector;
2817#ifdef CCISS_DEBUG
2818	printk(KERN_DEBUG "ciss: sector =%d nr_sectors=%d\n", (int)creq->sector,
2819	       (int)creq->nr_sectors);
2820#endif				/* CCISS_DEBUG */
2821
2822	sg_init_table(tmp_sg, MAXSGENTRIES);
2823	seg = blk_rq_map_sg(q, creq, tmp_sg);
2824
2825	/* get the DMA records for the setup */
2826	if (c->Request.Type.Direction == XFER_READ)
2827		dir = PCI_DMA_FROMDEVICE;
2828	else
2829		dir = PCI_DMA_TODEVICE;
2830
2831	for (i = 0; i < seg; i++) {
2832		c->SG[i].Len = tmp_sg[i].length;
2833		temp64.val = (__u64) pci_map_page(h->pdev, sg_page(&tmp_sg[i]),
2834						  tmp_sg[i].offset,
2835						  tmp_sg[i].length, dir);
2836		c->SG[i].Addr.lower = temp64.val32.lower;
2837		c->SG[i].Addr.upper = temp64.val32.upper;
2838		c->SG[i].Ext = 0;	// we are not chaining
2839	}
2840	/* track how many SG entries we are using */
2841	if (seg > h->maxSG)
2842		h->maxSG = seg;
2843
2844#ifdef CCISS_DEBUG
2845	printk(KERN_DEBUG "cciss: Submitting %d sectors in %d segments\n",
2846	       creq->nr_sectors, seg);
2847#endif				/* CCISS_DEBUG */
2848
2849	c->Header.SGList = c->Header.SGTotal = seg;
2850	if (likely(blk_fs_request(creq))) {
2851		if(h->cciss_read == CCISS_READ_10) {
2852			c->Request.CDB[1] = 0;
2853			c->Request.CDB[2] = (start_blk >> 24) & 0xff;	//MSB
2854			c->Request.CDB[3] = (start_blk >> 16) & 0xff;
2855			c->Request.CDB[4] = (start_blk >> 8) & 0xff;
2856			c->Request.CDB[5] = start_blk & 0xff;
2857			c->Request.CDB[6] = 0;	// (sect >> 24) & 0xff; MSB
2858			c->Request.CDB[7] = (creq->nr_sectors >> 8) & 0xff;
2859			c->Request.CDB[8] = creq->nr_sectors & 0xff;
2860			c->Request.CDB[9] = c->Request.CDB[11] = c->Request.CDB[12] = 0;
2861		} else {
2862			u32 upper32 = upper_32_bits(start_blk);
2863
2864			c->Request.CDBLen = 16;
2865			c->Request.CDB[1]= 0;
2866			c->Request.CDB[2]= (upper32 >> 24) & 0xff;	//MSB
2867			c->Request.CDB[3]= (upper32 >> 16) & 0xff;
2868			c->Request.CDB[4]= (upper32 >>  8) & 0xff;
2869			c->Request.CDB[5]= upper32 & 0xff;
2870			c->Request.CDB[6]= (start_blk >> 24) & 0xff;
2871			c->Request.CDB[7]= (start_blk >> 16) & 0xff;
2872			c->Request.CDB[8]= (start_blk >>  8) & 0xff;
2873			c->Request.CDB[9]= start_blk & 0xff;
2874			c->Request.CDB[10]= (creq->nr_sectors >>  24) & 0xff;
2875			c->Request.CDB[11]= (creq->nr_sectors >>  16) & 0xff;
2876			c->Request.CDB[12]= (creq->nr_sectors >>  8) & 0xff;
2877			c->Request.CDB[13]= creq->nr_sectors & 0xff;
2878			c->Request.CDB[14] = c->Request.CDB[15] = 0;
2879		}
2880	} else if (blk_pc_request(creq)) {
2881		c->Request.CDBLen = creq->cmd_len;
2882		memcpy(c->Request.CDB, creq->cmd, BLK_MAX_CDB);
2883	} else {
2884		printk(KERN_WARNING "cciss%d: bad request type %d\n", h->ctlr, creq->cmd_type);
2885		BUG();
2886	}
2887
2888	spin_lock_irq(q->queue_lock);
2889
2890	addQ(&(h->reqQ), c);
2891	h->Qdepth++;
2892	if (h->Qdepth > h->maxQsinceinit)
2893		h->maxQsinceinit = h->Qdepth;
2894
2895	goto queue;
2896full:
2897	blk_stop_queue(q);
2898startio:
2899	/* We will already have the driver lock here so not need
2900	 * to lock it.
2901	 */
2902	start_io(h);
2903}
2904
2905static inline unsigned long get_next_completion(ctlr_info_t *h)
2906{
2907#ifdef CONFIG_CISS_SCSI_TAPE
2908	/* Any rejects from sendcmd() lying around? Process them first */
2909	if (h->scsi_rejects.ncompletions == 0)
2910		return h->access.command_completed(h);
2911	else {
2912		struct sendcmd_reject_list *srl;
2913		int n;
2914		srl = &h->scsi_rejects;
2915		n = --srl->ncompletions;
2916		/* printk("cciss%d: processing saved reject\n", h->ctlr); */
2917		printk("p");
2918		return srl->complete[n];
2919	}
2920#else
2921	return h->access.command_completed(h);
2922#endif
2923}
2924
2925static inline int interrupt_pending(ctlr_info_t *h)
2926{
2927#ifdef CONFIG_CISS_SCSI_TAPE
2928	return (h->access.intr_pending(h)
2929		|| (h->scsi_rejects.ncompletions > 0));
2930#else
2931	return h->access.intr_pending(h);
2932#endif
2933}
2934
2935static inline long interrupt_not_for_us(ctlr_info_t *h)
2936{
2937#ifdef CONFIG_CISS_SCSI_TAPE
2938	return (((h->access.intr_pending(h) == 0) ||
2939		 (h->interrupts_enabled == 0))
2940		&& (h->scsi_rejects.ncompletions == 0));
2941#else
2942	return (((h->access.intr_pending(h) == 0) ||
2943		 (h->interrupts_enabled == 0)));
2944#endif
2945}
2946
2947static irqreturn_t do_cciss_intr(int irq, void *dev_id)
2948{
2949	ctlr_info_t *h = dev_id;
2950	CommandList_struct *c;
2951	unsigned long flags;
2952	__u32 a, a1, a2;
2953
2954	if (interrupt_not_for_us(h))
2955		return IRQ_NONE;
2956	/*
2957	 * If there are completed commands in the completion queue,
2958	 * we had better do something about it.
2959	 */
2960	spin_lock_irqsave(CCISS_LOCK(h->ctlr), flags);
2961	while (interrupt_pending(h)) {
2962		while ((a = get_next_completion(h)) != FIFO_EMPTY) {
2963			a1 = a;
2964			if ((a & 0x04)) {
2965				a2 = (a >> 3);
2966				if (a2 >= h->nr_cmds) {
2967					printk(KERN_WARNING
2968					       "cciss: controller cciss%d failed, stopping.\n",
2969					       h->ctlr);
2970					fail_all_cmds(h->ctlr);
2971					return IRQ_HANDLED;
2972				}
2973
2974				c = h->cmd_pool + a2;
2975				a = c->busaddr;
2976
2977			} else {
2978				a &= ~3;
2979				if ((c = h->cmpQ) == NULL) {
2980					printk(KERN_WARNING
2981					       "cciss: Completion of %08x ignored\n",
2982					       a1);
2983					continue;
2984				}
2985				while (c->busaddr != a) {
2986					c = c->next;
2987					if (c == h->cmpQ)
2988						break;
2989				}
2990			}
2991			/*
2992			 * If we've found the command, take it off the
2993			 * completion Q and free it
2994			 */
2995			if (c->busaddr == a) {
2996				removeQ(&h->cmpQ, c);
2997				if (c->cmd_type == CMD_RWREQ) {
2998					complete_command(h, c, 0);
2999				} else if (c->cmd_type == CMD_IOCTL_PEND) {
3000					complete(c->waiting);
3001				}
3002#				ifdef CONFIG_CISS_SCSI_TAPE
3003				else if (c->cmd_type == CMD_SCSI)
3004					complete_scsi_command(c, 0, a1);
3005#				endif
3006				continue;
3007			}
3008		}
3009	}
3010
3011	spin_unlock_irqrestore(CCISS_LOCK(h->ctlr), flags);
3012	return IRQ_HANDLED;
3013}
3014
3015/*
3016 *  We cannot read the structure directly, for portability we must use
3017 *   the io functions.
3018 *   This is for debug only.
3019 */
3020#ifdef CCISS_DEBUG
3021static void print_cfg_table(CfgTable_struct *tb)
3022{
3023	int i;
3024	char temp_name[17];
3025
3026	printk("Controller Configuration information\n");
3027	printk("------------------------------------\n");
3028	for (i = 0; i < 4; i++)
3029		temp_name[i] = readb(&(tb->Signature[i]));
3030	temp_name[4] = '\0';
3031	printk("   Signature = %s\n", temp_name);
3032	printk("   Spec Number = %d\n", readl(&(tb->SpecValence)));
3033	printk("   Transport methods supported = 0x%x\n",
3034	       readl(&(tb->TransportSupport)));
3035	printk("   Transport methods active = 0x%x\n",
3036	       readl(&(tb->TransportActive)));
3037	printk("   Requested transport Method = 0x%x\n",
3038	       readl(&(tb->HostWrite.TransportRequest)));
3039	printk("   Coalesce Interrupt Delay = 0x%x\n",
3040	       readl(&(tb->HostWrite.CoalIntDelay)));
3041	printk("   Coalesce Interrupt Count = 0x%x\n",
3042	       readl(&(tb->HostWrite.CoalIntCount)));
3043	printk("   Max outstanding commands = 0x%d\n",
3044	       readl(&(tb->CmdsOutMax)));
3045	printk("   Bus Types = 0x%x\n", readl(&(tb->BusTypes)));
3046	for (i = 0; i < 16; i++)
3047		temp_name[i] = readb(&(tb->ServerName[i]));
3048	temp_name[16] = '\0';
3049	printk("   Server Name = %s\n", temp_name);
3050	printk("   Heartbeat Counter = 0x%x\n\n\n", readl(&(tb->HeartBeat)));
3051}
3052#endif				/* CCISS_DEBUG */
3053
3054static int find_PCI_BAR_index(struct pci_dev *pdev, unsigned long pci_bar_addr)
3055{
3056	int i, offset, mem_type, bar_type;
3057	if (pci_bar_addr == PCI_BASE_ADDRESS_0)	/* looking for BAR zero? */
3058		return 0;
3059	offset = 0;
3060	for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
3061		bar_type = pci_resource_flags(pdev, i) & PCI_BASE_ADDRESS_SPACE;
3062		if (bar_type == PCI_BASE_ADDRESS_SPACE_IO)
3063			offset += 4;
3064		else {
3065			mem_type = pci_resource_flags(pdev, i) &
3066			    PCI_BASE_ADDRESS_MEM_TYPE_MASK;
3067			switch (mem_type) {
3068			case PCI_BASE_ADDRESS_MEM_TYPE_32:
3069			case PCI_BASE_ADDRESS_MEM_TYPE_1M:
3070				offset += 4;	/* 32 bit */
3071				break;
3072			case PCI_BASE_ADDRESS_MEM_TYPE_64:
3073				offset += 8;
3074				break;
3075			default:	/* reserved in PCI 2.2 */
3076				printk(KERN_WARNING
3077				       "Base address is invalid\n");
3078				return -1;
3079				break;
3080			}
3081		}
3082		if (offset == pci_bar_addr - PCI_BASE_ADDRESS_0)
3083			return i + 1;
3084	}
3085	return -1;
3086}
3087
3088/* If MSI/MSI-X is supported by the kernel we will try to enable it on
3089 * controllers that are capable. If not, we use IO-APIC mode.
3090 */
3091
3092static void __devinit cciss_interrupt_mode(ctlr_info_t *c,
3093					   struct pci_dev *pdev, __u32 board_id)
3094{
3095#ifdef CONFIG_PCI_MSI
3096	int err;
3097	struct msix_entry cciss_msix_entries[4] = { {0, 0}, {0, 1},
3098	{0, 2}, {0, 3}
3099	};
3100
3101	/* Some boards advertise MSI but don't really support it */
3102	if ((board_id == 0x40700E11) ||
3103	    (board_id == 0x40800E11) ||
3104	    (board_id == 0x40820E11) || (board_id == 0x40830E11))
3105		goto default_int_mode;
3106
3107	if (pci_find_capability(pdev, PCI_CAP_ID_MSIX)) {
3108		err = pci_enable_msix(pdev, cciss_msix_entries, 4);
3109		if (!err) {
3110			c->intr[0] = cciss_msix_entries[0].vector;
3111			c->intr[1] = cciss_msix_entries[1].vector;
3112			c->intr[2] = cciss_msix_entries[2].vector;
3113			c->intr[3] = cciss_msix_entries[3].vector;
3114			c->msix_vector = 1;
3115			return;
3116		}
3117		if (err > 0) {
3118			printk(KERN_WARNING "cciss: only %d MSI-X vectors "
3119			       "available\n", err);
3120			goto default_int_mode;
3121		} else {
3122			printk(KERN_WARNING "cciss: MSI-X init failed %d\n",
3123			       err);
3124			goto default_int_mode;
3125		}
3126	}
3127	if (pci_find_capability(pdev, PCI_CAP_ID_MSI)) {
3128		if (!pci_enable_msi(pdev)) {
3129			c->msi_vector = 1;
3130		} else {
3131			printk(KERN_WARNING "cciss: MSI init failed\n");
3132		}
3133	}
3134default_int_mode:
3135#endif				/* CONFIG_PCI_MSI */
3136	/* if we get here we're going to use the default interrupt mode */
3137	c->intr[SIMPLE_MODE_INT] = pdev->irq;
3138	return;
3139}
3140
3141static int __devinit cciss_pci_init(ctlr_info_t *c, struct pci_dev *pdev)
3142{
3143	ushort subsystem_vendor_id, subsystem_device_id, command;
3144	__u32 board_id, scratchpad = 0;
3145	__u64 cfg_offset;
3146	__u32 cfg_base_addr;
3147	__u64 cfg_base_addr_index;
3148	int i, err;
3149
3150	/* check to see if controller has been disabled */
3151	/* BEFORE trying to enable it */
3152	(void)pci_read_config_word(pdev, PCI_COMMAND, &command);
3153	if (!(command & 0x02)) {
3154		printk(KERN_WARNING
3155		       "cciss: controller appears to be disabled\n");
3156		return -ENODEV;
3157	}
3158
3159	err = pci_enable_device(pdev);
3160	if (err) {
3161		printk(KERN_ERR "cciss: Unable to Enable PCI device\n");
3162		return err;
3163	}
3164
3165	err = pci_request_regions(pdev, "cciss");
3166	if (err) {
3167		printk(KERN_ERR "cciss: Cannot obtain PCI resources, "
3168		       "aborting\n");
3169		return err;
3170	}
3171
3172	subsystem_vendor_id = pdev->subsystem_vendor;
3173	subsystem_device_id = pdev->subsystem_device;
3174	board_id = (((__u32) (subsystem_device_id << 16) & 0xffff0000) |
3175		    subsystem_vendor_id);
3176
3177#ifdef CCISS_DEBUG
3178	printk("command = %x\n", command);
3179	printk("irq = %x\n", pdev->irq);
3180	printk("board_id = %x\n", board_id);
3181#endif				/* CCISS_DEBUG */
3182
3183/* If the kernel supports MSI/MSI-X we will try to enable that functionality,
3184 * else we use the IO-APIC interrupt assigned to us by system ROM.
3185 */
3186	cciss_interrupt_mode(c, pdev, board_id);
3187
3188	/*
3189	 * Memory base addr is first addr , the second points to the config
3190	 *   table
3191	 */
3192
3193	c->paddr = pci_resource_start(pdev, 0);	/* addressing mode bits already removed */
3194#ifdef CCISS_DEBUG
3195	printk("address 0 = %x\n", c->paddr);
3196#endif				/* CCISS_DEBUG */
3197	c->vaddr = remap_pci_mem(c->paddr, 0x250);
3198
3199	/* Wait for the board to become ready.  (PCI hotplug needs this.)
3200	 * We poll for up to 120 secs, once per 100ms. */
3201	for (i = 0; i < 1200; i++) {
3202		scratchpad = readl(c->vaddr + SA5_SCRATCHPAD_OFFSET);
3203		if (scratchpad == CCISS_FIRMWARE_READY)
3204			break;
3205		set_current_state(TASK_INTERRUPTIBLE);
3206		schedule_timeout(HZ / 10);	/* wait 100ms */
3207	}
3208	if (scratchpad != CCISS_FIRMWARE_READY) {
3209		printk(KERN_WARNING "cciss: Board not ready.  Timed out.\n");
3210		err = -ENODEV;
3211		goto err_out_free_res;
3212	}
3213
3214	/* get the address index number */
3215	cfg_base_addr = readl(c->vaddr + SA5_CTCFG_OFFSET);
3216	cfg_base_addr &= (__u32) 0x0000ffff;
3217#ifdef CCISS_DEBUG
3218	printk("cfg base address = %x\n", cfg_base_addr);
3219#endif				/* CCISS_DEBUG */
3220	cfg_base_addr_index = find_PCI_BAR_index(pdev, cfg_base_addr);
3221#ifdef CCISS_DEBUG
3222	printk("cfg base address index = %x\n", cfg_base_addr_index);
3223#endif				/* CCISS_DEBUG */
3224	if (cfg_base_addr_index == -1) {
3225		printk(KERN_WARNING "cciss: Cannot find cfg_base_addr_index\n");
3226		err = -ENODEV;
3227		goto err_out_free_res;
3228	}
3229
3230	cfg_offset = readl(c->vaddr + SA5_CTMEM_OFFSET);
3231#ifdef CCISS_DEBUG
3232	printk("cfg offset = %x\n", cfg_offset);
3233#endif				/* CCISS_DEBUG */
3234	c->cfgtable = remap_pci_mem(pci_resource_start(pdev,
3235						       cfg_base_addr_index) +
3236				    cfg_offset, sizeof(CfgTable_struct));
3237	c->board_id = board_id;
3238
3239#ifdef CCISS_DEBUG
3240	print_cfg_table(c->cfgtable);
3241#endif				/* CCISS_DEBUG */
3242
3243	/* Some controllers support Zero Memory Raid (ZMR).
3244	 * When configured in ZMR mode the number of supported
3245	 * commands drops to 64. So instead of just setting an
3246	 * arbitrary value we make the driver a little smarter.
3247	 * We read the config table to tell us how many commands
3248	 * are supported on the controller then subtract 4 to
3249	 * leave a little room for ioctl calls.
3250	 */
3251	c->max_commands = readl(&(c->cfgtable->CmdsOutMax));
3252	for (i = 0; i < ARRAY_SIZE(products); i++) {
3253		if (board_id == products[i].board_id) {
3254			c->product_name = products[i].product_name;
3255			c->access = *(products[i].access);
3256			c->nr_cmds = c->max_commands - 4;
3257			break;
3258		}
3259	}
3260	if ((readb(&c->cfgtable->Signature[0]) != 'C') ||
3261	    (readb(&c->cfgtable->Signature[1]) != 'I') ||
3262	    (readb(&c->cfgtable->Signature[2]) != 'S') ||
3263	    (readb(&c->cfgtable->Signature[3]) != 'S')) {
3264		printk("Does not appear to be a valid CISS config table\n");
3265		err = -ENODEV;
3266		goto err_out_free_res;
3267	}
3268	/* We didn't find the controller in our list. We know the
3269	 * signature is valid. If it's an HP device let's try to
3270	 * bind to the device and fire it up. Otherwise we bail.
3271	 */
3272	if (i == ARRAY_SIZE(products)) {
3273		if (subsystem_vendor_id == PCI_VENDOR_ID_HP) {
3274			c->product_name = products[i-1].product_name;
3275			c->access = *(products[i-1].access);
3276			c->nr_cmds = c->max_commands - 4;
3277			printk(KERN_WARNING "cciss: This is an unknown "
3278				"Smart Array controller.\n"
3279				"cciss: Please update to the latest driver "
3280				"available from www.hp.com.\n");
3281		} else {
3282			printk(KERN_WARNING "cciss: Sorry, I don't know how"
3283				" to access the Smart Array controller %08lx\n"
3284					, (unsigned long)board_id);
3285			err = -ENODEV;
3286			goto err_out_free_res;
3287		}
3288	}
3289#ifdef CONFIG_X86
3290	{
3291		/* Need to enable prefetch in the SCSI core for 6400 in x86 */
3292		__u32 prefetch;
3293		prefetch = readl(&(c->cfgtable->SCSI_Prefetch));
3294		prefetch |= 0x100;
3295		writel(prefetch, &(c->cfgtable->SCSI_Prefetch));
3296	}
3297#endif
3298
3299	/* Disabling DMA prefetch and refetch for the P600.
3300	 * An ASIC bug may result in accesses to invalid memory addresses.
3301	 * We've disabled prefetch for some time now. Testing with XEN
3302	 * kernels revealed a bug in the refetch if dom0 resides on a P600.
3303	 */
3304	if(board_id == 0x3225103C) {
3305		__u32 dma_prefetch;
3306		__u32 dma_refetch;
3307		dma_prefetch = readl(c->vaddr + I2O_DMA1_CFG);
3308		dma_prefetch |= 0x8000;
3309		writel(dma_prefetch, c->vaddr + I2O_DMA1_CFG);
3310		pci_read_config_dword(pdev, PCI_COMMAND_PARITY, &dma_refetch);
3311		dma_refetch |= 0x1;
3312		pci_write_config_dword(pdev, PCI_COMMAND_PARITY, dma_refetch);
3313	}
3314
3315#ifdef CCISS_DEBUG
3316	printk("Trying to put board into Simple mode\n");
3317#endif				/* CCISS_DEBUG */
3318	c->max_commands = readl(&(c->cfgtable->CmdsOutMax));
3319	/* Update the field, and then ring the doorbell */
3320	writel(CFGTBL_Trans_Simple, &(c->cfgtable->HostWrite.TransportRequest));
3321	writel(CFGTBL_ChangeReq, c->vaddr + SA5_DOORBELL);
3322
3323	/* under certain very rare conditions, this can take awhile.
3324	 * (e.g.: hot replace a failed 144GB drive in a RAID 5 set right
3325	 * as we enter this code.) */
3326	for (i = 0; i < MAX_CONFIG_WAIT; i++) {
3327		if (!(readl(c->vaddr + SA5_DOORBELL) & CFGTBL_ChangeReq))
3328			break;
3329		/* delay and try again */
3330		set_current_state(TASK_INTERRUPTIBLE);
3331		schedule_timeout(10);
3332	}
3333
3334#ifdef CCISS_DEBUG
3335	printk(KERN_DEBUG "I counter got to %d %x\n", i,
3336	       readl(c->vaddr + SA5_DOORBELL));
3337#endif				/* CCISS_DEBUG */
3338#ifdef CCISS_DEBUG
3339	print_cfg_table(c->cfgtable);
3340#endif				/* CCISS_DEBUG */
3341
3342	if (!(readl(&(c->cfgtable->TransportActive)) & CFGTBL_Trans_Simple)) {
3343		printk(KERN_WARNING "cciss: unable to get board into"
3344		       " simple mode\n");
3345		err = -ENODEV;
3346		goto err_out_free_res;
3347	}
3348	return 0;
3349
3350err_out_free_res:
3351	/*
3352	 * Deliberately omit pci_disable_device(): it does something nasty to
3353	 * Smart Array controllers that pci_enable_device does not undo
3354	 */
3355	pci_release_regions(pdev);
3356	return err;
3357}
3358
3359/* Function to find the first free pointer into our hba[] array
3360 * Returns -1 if no free entries are left.
3361 */
3362static int alloc_cciss_hba(void)
3363{
3364	int i;
3365
3366	for (i = 0; i < MAX_CTLR; i++) {
3367		if (!hba[i]) {
3368			ctlr_info_t *p;
3369
3370			p = kzalloc(sizeof(ctlr_info_t), GFP_KERNEL);
3371			if (!p)
3372				goto Enomem;
3373			hba[i] = p;
3374			return i;
3375		}
3376	}
3377	printk(KERN_WARNING "cciss: This driver supports a maximum"
3378	       " of %d controllers.\n", MAX_CTLR);
3379	return -1;
3380Enomem:
3381	printk(KERN_ERR "cciss: out of memory.\n");
3382	return -1;
3383}
3384
3385static void free_hba(int i)
3386{
3387	ctlr_info_t *p = hba[i];
3388	int n;
3389
3390	hba[i] = NULL;
3391	for (n = 0; n < CISS_MAX_LUN; n++)
3392		put_disk(p->gendisk[n]);
3393	kfree(p);
3394}
3395
3396/*
3397 *  This is it.  Find all the controllers and register them.  I really hate
3398 *  stealing all these major device numbers.
3399 *  returns the number of block devices registered.
3400 */
3401static int __devinit cciss_init_one(struct pci_dev *pdev,
3402				    const struct pci_device_id *ent)
3403{
3404	int i;
3405	int j = 0;
3406	int rc;
3407	int dac;
3408
3409	i = alloc_cciss_hba();
3410	if (i < 0)
3411		return -1;
3412
3413	hba[i]->busy_initializing = 1;
3414
3415	if (cciss_pci_init(hba[i], pdev) != 0)
3416		goto clean1;
3417
3418	sprintf(hba[i]->devname, "cciss%d", i);
3419	hba[i]->ctlr = i;
3420	hba[i]->pdev = pdev;
3421
3422	/* configure PCI DMA stuff */
3423	if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK))
3424		dac = 1;
3425	else if (!pci_set_dma_mask(pdev, DMA_32BIT_MASK))
3426		dac = 0;
3427	else {
3428		printk(KERN_ERR "cciss: no suitable DMA available\n");
3429		goto clean1;
3430	}
3431
3432	/*
3433	 * register with the major number, or get a dynamic major number
3434	 * by passing 0 as argument.  This is done for greater than
3435	 * 8 controller support.
3436	 */
3437	if (i < MAX_CTLR_ORIG)
3438		hba[i]->major = COMPAQ_CISS_MAJOR + i;
3439	rc = register_blkdev(hba[i]->major, hba[i]->devname);
3440	if (rc == -EBUSY || rc == -EINVAL) {
3441		printk(KERN_ERR
3442		       "cciss:  Unable to get major number %d for %s "
3443		       "on hba %d\n", hba[i]->major, hba[i]->devname, i);
3444		goto clean1;
3445	} else {
3446		if (i >= MAX_CTLR_ORIG)
3447			hba[i]->major = rc;
3448	}
3449
3450	/* make sure the board interrupts are off */
3451	hba[i]->access.set_intr_mask(hba[i], CCISS_INTR_OFF);
3452	if (request_irq(hba[i]->intr[SIMPLE_MODE_INT], do_cciss_intr,
3453			IRQF_DISABLED | IRQF_SHARED, hba[i]->devname, hba[i])) {
3454		printk(KERN_ERR "cciss: Unable to get irq %d for %s\n",
3455		       hba[i]->intr[SIMPLE_MODE_INT], hba[i]->devname);
3456		goto clean2;
3457	}
3458
3459	printk(KERN_INFO "%s: <0x%x> at PCI %s IRQ %d%s using DAC\n",
3460	       hba[i]->devname, pdev->device, pci_name(pdev),
3461	       hba[i]->intr[SIMPLE_MODE_INT], dac ? "" : " not");
3462
3463	hba[i]->cmd_pool_bits =
3464	    kmalloc(DIV_ROUND_UP(hba[i]->nr_cmds, BITS_PER_LONG)
3465			* sizeof(unsigned long), GFP_KERNEL);
3466	hba[i]->cmd_pool = (CommandList_struct *)
3467	    pci_alloc_consistent(hba[i]->pdev,
3468		    hba[i]->nr_cmds * sizeof(CommandList_struct),
3469		    &(hba[i]->cmd_pool_dhandle));
3470	hba[i]->errinfo_pool = (ErrorInfo_struct *)
3471	    pci_alloc_consistent(hba[i]->pdev,
3472		    hba[i]->nr_cmds * sizeof(ErrorInfo_struct),
3473		    &(hba[i]->errinfo_pool_dhandle));
3474	if ((hba[i]->cmd_pool_bits == NULL)
3475	    || (hba[i]->cmd_pool == NULL)
3476	    || (hba[i]->errinfo_pool == NULL)) {
3477		printk(KERN_ERR "cciss: out of memory");
3478		goto clean4;
3479	}
3480#ifdef CONFIG_CISS_SCSI_TAPE
3481	hba[i]->scsi_rejects.complete =
3482	    kmalloc(sizeof(hba[i]->scsi_rejects.complete[0]) *
3483		    (hba[i]->nr_cmds + 5), GFP_KERNEL);
3484	if (hba[i]->scsi_rejects.complete == NULL) {
3485		printk(KERN_ERR "cciss: out of memory");
3486		goto clean4;
3487	}
3488#endif
3489	spin_lock_init(&hba[i]->lock);
3490
3491	/* Initialize the pdev driver private data.
3492	   have it point to hba[i].  */
3493	pci_set_drvdata(pdev, hba[i]);
3494	/* command and error info recs zeroed out before
3495	   they are used */
3496	memset(hba[i]->cmd_pool_bits, 0,
3497	       DIV_ROUND_UP(hba[i]->nr_cmds, BITS_PER_LONG)
3498			* sizeof(unsigned long));
3499
3500	hba[i]->num_luns = 0;
3501	hba[i]->highest_lun = -1;
3502	for (j = 0; j < CISS_MAX_LUN; j++) {
3503		hba[i]->drv[j].raid_level = -1;
3504		hba[i]->drv[j].queue = NULL;
3505		hba[i]->gendisk[j] = NULL;
3506	}
3507
3508	cciss_scsi_setup(i);
3509
3510	/* Turn the interrupts on so we can service requests */
3511	hba[i]->access.set_intr_mask(hba[i], CCISS_INTR_ON);
3512
3513	cciss_procinit(i);
3514
3515	hba[i]->cciss_max_sectors = 2048;
3516
3517	hba[i]->busy_initializing = 0;
3518
3519	rebuild_lun_table(hba[i], 1);
3520	return 1;
3521
3522clean4:
3523#ifdef CONFIG_CISS_SCSI_TAPE
3524	kfree(hba[i]->scsi_rejects.complete);
3525#endif
3526	kfree(hba[i]->cmd_pool_bits);
3527	if (hba[i]->cmd_pool)
3528		pci_free_consistent(hba[i]->pdev,
3529				    hba[i]->nr_cmds * sizeof(CommandList_struct),
3530				    hba[i]->cmd_pool, hba[i]->cmd_pool_dhandle);
3531	if (hba[i]->errinfo_pool)
3532		pci_free_consistent(hba[i]->pdev,
3533				    hba[i]->nr_cmds * sizeof(ErrorInfo_struct),
3534				    hba[i]->errinfo_pool,
3535				    hba[i]->errinfo_pool_dhandle);
3536	free_irq(hba[i]->intr[SIMPLE_MODE_INT], hba[i]);
3537clean2:
3538	unregister_blkdev(hba[i]->major, hba[i]->devname);
3539clean1:
3540	hba[i]->busy_initializing = 0;
3541	/* cleanup any queues that may have been initialized */
3542	for (j=0; j <= hba[i]->highest_lun; j++){
3543		drive_info_struct *drv = &(hba[i]->drv[j]);
3544		if (drv->queue)
3545			blk_cleanup_queue(drv->queue);
3546	}
3547	/*
3548	 * Deliberately omit pci_disable_device(): it does something nasty to
3549	 * Smart Array controllers that pci_enable_device does not undo
3550	 */
3551	pci_release_regions(pdev);
3552	pci_set_drvdata(pdev, NULL);
3553	free_hba(i);
3554	return -1;
3555}
3556
3557static void cciss_shutdown(struct pci_dev *pdev)
3558{
3559	ctlr_info_t *tmp_ptr;
3560	int i;
3561	char flush_buf[4];
3562	int return_code;
3563
3564	tmp_ptr = pci_get_drvdata(pdev);
3565	if (tmp_ptr == NULL)
3566		return;
3567	i = tmp_ptr->ctlr;
3568	if (hba[i] == NULL)
3569		return;
3570
3571	/* Turn board interrupts off  and send the flush cache command */
3572	/* sendcmd will turn off interrupt, and send the flush...
3573	 * To write all data in the battery backed cache to disks */
3574	memset(flush_buf, 0, 4);
3575	return_code = sendcmd(CCISS_CACHE_FLUSH, i, flush_buf, 4, 0, 0, 0, NULL,
3576			      TYPE_CMD);
3577	if (return_code == IO_OK) {
3578		printk(KERN_INFO "Completed flushing cache on controller %d\n", i);
3579	} else {
3580		printk(KERN_WARNING "Error flushing cache on controller %d\n", i);
3581	}
3582	free_irq(hba[i]->intr[2], hba[i]);
3583}
3584
3585static void __devexit cciss_remove_one(struct pci_dev *pdev)
3586{
3587	ctlr_info_t *tmp_ptr;
3588	int i, j;
3589
3590	if (pci_get_drvdata(pdev) == NULL) {
3591		printk(KERN_ERR "cciss: Unable to remove device \n");
3592		return;
3593	}
3594	tmp_ptr = pci_get_drvdata(pdev);
3595	i = tmp_ptr->ctlr;
3596	if (hba[i] == NULL) {
3597		printk(KERN_ERR "cciss: device appears to "
3598		       "already be removed \n");
3599		return;
3600	}
3601
3602	remove_proc_entry(hba[i]->devname, proc_cciss);
3603	unregister_blkdev(hba[i]->major, hba[i]->devname);
3604
3605	/* remove it from the disk list */
3606	for (j = 0; j < CISS_MAX_LUN; j++) {
3607		struct gendisk *disk = hba[i]->gendisk[j];
3608		if (disk) {
3609			struct request_queue *q = disk->queue;
3610
3611			if (disk->flags & GENHD_FL_UP)
3612				del_gendisk(disk);
3613			if (q)
3614				blk_cleanup_queue(q);
3615		}
3616	}
3617
3618#ifdef CONFIG_CISS_SCSI_TAPE
3619	cciss_unregister_scsi(i);	/* unhook from SCSI subsystem */
3620#endif
3621
3622	cciss_shutdown(pdev);
3623
3624#ifdef CONFIG_PCI_MSI
3625	if (hba[i]->msix_vector)
3626		pci_disable_msix(hba[i]->pdev);
3627	else if (hba[i]->msi_vector)
3628		pci_disable_msi(hba[i]->pdev);
3629#endif				/* CONFIG_PCI_MSI */
3630
3631	iounmap(hba[i]->vaddr);
3632
3633	pci_free_consistent(hba[i]->pdev, hba[i]->nr_cmds * sizeof(CommandList_struct),
3634			    hba[i]->cmd_pool, hba[i]->cmd_pool_dhandle);
3635	pci_free_consistent(hba[i]->pdev, hba[i]->nr_cmds * sizeof(ErrorInfo_struct),
3636			    hba[i]->errinfo_pool, hba[i]->errinfo_pool_dhandle);
3637	kfree(hba[i]->cmd_pool_bits);
3638#ifdef CONFIG_CISS_SCSI_TAPE
3639	kfree(hba[i]->scsi_rejects.complete);
3640#endif
3641	/*
3642	 * Deliberately omit pci_disable_device(): it does something nasty to
3643	 * Smart Array controllers that pci_enable_device does not undo
3644	 */
3645	pci_release_regions(pdev);
3646	pci_set_drvdata(pdev, NULL);
3647	free_hba(i);
3648}
3649
3650static struct pci_driver cciss_pci_driver = {
3651	.name = "cciss",
3652	.probe = cciss_init_one,
3653	.remove = __devexit_p(cciss_remove_one),
3654	.id_table = cciss_pci_device_id,	/* id_table */
3655	.shutdown = cciss_shutdown,
3656};
3657
3658/*
3659 *  This is it.  Register the PCI driver information for the cards we control
3660 *  the OS will call our registered routines when it finds one of our cards.
3661 */
3662static int __init cciss_init(void)
3663{
3664	printk(KERN_INFO DRIVER_NAME "\n");
3665
3666	/* Register for our PCI devices */
3667	return pci_register_driver(&cciss_pci_driver);
3668}
3669
3670static void __exit cciss_cleanup(void)
3671{
3672	int i;
3673
3674	pci_unregister_driver(&cciss_pci_driver);
3675	/* double check that all controller entrys have been removed */
3676	for (i = 0; i < MAX_CTLR; i++) {
3677		if (hba[i] != NULL) {
3678			printk(KERN_WARNING "cciss: had to remove"
3679			       " controller %d\n", i);
3680			cciss_remove_one(hba[i]->pdev);
3681		}
3682	}
3683	remove_proc_entry("driver/cciss", NULL);
3684}
3685
3686static void fail_all_cmds(unsigned long ctlr)
3687{
3688	/* If we get here, the board is apparently dead. */
3689	ctlr_info_t *h = hba[ctlr];
3690	CommandList_struct *c;
3691	unsigned long flags;
3692
3693	printk(KERN_WARNING "cciss%d: controller not responding.\n", h->ctlr);
3694	h->alive = 0;		/* the controller apparently died... */
3695
3696	spin_lock_irqsave(CCISS_LOCK(ctlr), flags);
3697
3698	pci_disable_device(h->pdev);	/* Make sure it is really dead. */
3699
3700	/* move everything off the request queue onto the completed queue */
3701	while ((c = h->reqQ) != NULL) {
3702		removeQ(&(h->reqQ), c);
3703		h->Qdepth--;
3704		addQ(&(h->cmpQ), c);
3705	}
3706
3707	/* Now, fail everything on the completed queue with a HW error */
3708	while ((c = h->cmpQ) != NULL) {
3709		removeQ(&h->cmpQ, c);
3710		c->err_info->CommandStatus = CMD_HARDWARE_ERR;
3711		if (c->cmd_type == CMD_RWREQ) {
3712			complete_command(h, c, 0);
3713		} else if (c->cmd_type == CMD_IOCTL_PEND)
3714			complete(c->waiting);
3715#ifdef CONFIG_CISS_SCSI_TAPE
3716		else if (c->cmd_type == CMD_SCSI)
3717			complete_scsi_command(c, 0, 0);
3718#endif
3719	}
3720	spin_unlock_irqrestore(CCISS_LOCK(ctlr), flags);
3721	return;
3722}
3723
3724module_init(cciss_init);
3725module_exit(cciss_cleanup);
3726