]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blame - drivers/scsi/dpt_i2o.c
Merge branch 'for-linus-4.10' of git://git.kernel.org/pub/scm/linux/kernel/git/mason...
[mirror_ubuntu-zesty-kernel.git] / drivers / scsi / dpt_i2o.c
CommitLineData
1da177e4
LT
1/***************************************************************************
2 dpti.c - description
3 -------------------
4 begin : Thu Sep 7 2000
5 copyright : (C) 2000 by Adaptec
6
7 July 30, 2001 First version being submitted
8 for inclusion in the kernel. V2.4
9
10 See Documentation/scsi/dpti.txt for history, notes, license info
11 and credits
12 ***************************************************************************/
13
14/***************************************************************************
15 * *
16 * This program is free software; you can redistribute it and/or modify *
17 * it under the terms of the GNU General Public License as published by *
18 * the Free Software Foundation; either version 2 of the License, or *
19 * (at your option) any later version. *
20 * *
21 ***************************************************************************/
22/***************************************************************************
23 * Sat Dec 20 2003 Go Taniguchi <go@turbolinux.co.jp>
24 - Support 2.6 kernel and DMA-mapping
25 - ioctl fix for raid tools
26 - use schedule_timeout in long long loop
27 **************************************************************************/
28
29/*#define DEBUG 1 */
30/*#define UARTDELAY 1 */
31
1da177e4
LT
32#include <linux/module.h>
33
34MODULE_AUTHOR("Deanna Bonds, with _lots_ of help from Mark Salyzyn");
35MODULE_DESCRIPTION("Adaptec I2O RAID Driver");
36
37////////////////////////////////////////////////////////////////
38
39#include <linux/ioctl.h> /* For SCSI-Passthrough */
7c0f6ba6 40#include <linux/uaccess.h>
1da177e4
LT
41
42#include <linux/stat.h>
43#include <linux/slab.h> /* for kmalloc() */
1da177e4
LT
44#include <linux/pci.h> /* for PCI support */
45#include <linux/proc_fs.h>
46#include <linux/blkdev.h>
47#include <linux/delay.h> /* for udelay */
48#include <linux/interrupt.h>
49#include <linux/kernel.h> /* for printk */
50#include <linux/sched.h>
51#include <linux/reboot.h>
52#include <linux/spinlock.h>
910638ae 53#include <linux/dma-mapping.h>
1da177e4
LT
54
55#include <linux/timer.h>
56#include <linux/string.h>
57#include <linux/ioport.h>
0b950672 58#include <linux/mutex.h>
1da177e4
LT
59
60#include <asm/processor.h> /* for boot_cpu_data */
61#include <asm/pgtable.h>
62#include <asm/io.h> /* for virt_to_bus, etc. */
63
64#include <scsi/scsi.h>
65#include <scsi/scsi_cmnd.h>
66#include <scsi/scsi_device.h>
67#include <scsi/scsi_host.h>
68#include <scsi/scsi_tcq.h>
69
70#include "dpt/dptsig.h"
71#include "dpti.h"
72
73/*============================================================================
74 * Create a binary signature - this is read by dptsig
75 * Needed for our management apps
76 *============================================================================
77 */
c45d15d2 78static DEFINE_MUTEX(adpt_mutex);
1da177e4
LT
79static dpt_sig_S DPTI_sig = {
80 {'d', 'P', 't', 'S', 'i', 'G'}, SIG_VERSION,
81#ifdef __i386__
82 PROC_INTEL, PROC_386 | PROC_486 | PROC_PENTIUM | PROC_SEXIUM,
83#elif defined(__ia64__)
84 PROC_INTEL, PROC_IA64,
85#elif defined(__sparc__)
86 PROC_ULTRASPARC, PROC_ULTRASPARC,
87#elif defined(__alpha__)
88 PROC_ALPHA, PROC_ALPHA,
89#else
90 (-1),(-1),
91#endif
92 FT_HBADRVR, 0, OEM_DPT, OS_LINUX, CAP_OVERLAP, DEV_ALL,
93 ADF_ALL_SC5, 0, 0, DPT_VERSION, DPT_REVISION, DPT_SUBREVISION,
94 DPT_MONTH, DPT_DAY, DPT_YEAR, "Adaptec Linux I2O RAID Driver"
95};
96
97
98
99
100/*============================================================================
101 * Globals
102 *============================================================================
103 */
104
0b950672 105static DEFINE_MUTEX(adpt_configuration_lock);
1da177e4 106
67af2b06
MS
107static struct i2o_sys_tbl *sys_tbl;
108static dma_addr_t sys_tbl_pa;
109static int sys_tbl_ind;
110static int sys_tbl_len;
1da177e4 111
1da177e4
LT
112static adpt_hba* hba_chain = NULL;
113static int hba_count = 0;
114
1ed43910
MS
115static struct class *adpt_sysfs_class;
116
f4927c45 117static long adpt_unlocked_ioctl(struct file *, unsigned int, unsigned long);
62ac5aed
MS
118#ifdef CONFIG_COMPAT
119static long compat_adpt_ioctl(struct file *, unsigned int, unsigned long);
120#endif
121
00977a59 122static const struct file_operations adpt_fops = {
f4927c45 123 .unlocked_ioctl = adpt_unlocked_ioctl,
1da177e4 124 .open = adpt_open,
62ac5aed
MS
125 .release = adpt_close,
126#ifdef CONFIG_COMPAT
127 .compat_ioctl = compat_adpt_ioctl,
1da177e4 128#endif
6038f373 129 .llseek = noop_llseek,
1da177e4 130};
1da177e4
LT
131
132/* Structures and definitions for synchronous message posting.
133 * See adpt_i2o_post_wait() for description
134 * */
135struct adpt_i2o_post_wait_data
136{
137 int status;
138 u32 id;
139 adpt_wait_queue_head_t *wq;
140 struct adpt_i2o_post_wait_data *next;
141};
142
143static struct adpt_i2o_post_wait_data *adpt_post_wait_queue = NULL;
144static u32 adpt_post_wait_id = 0;
145static DEFINE_SPINLOCK(adpt_post_wait_lock);
146
147
148/*============================================================================
149 * Functions
150 *============================================================================
151 */
152
62ac5aed
MS
153static inline int dpt_dma64(adpt_hba *pHba)
154{
155 return (sizeof(dma_addr_t) > 4 && (pHba)->dma64);
156}
157
67af2b06
MS
158static inline u32 dma_high(dma_addr_t addr)
159{
160 return upper_32_bits(addr);
161}
162
163static inline u32 dma_low(dma_addr_t addr)
164{
165 return (u32)addr;
166}
167
1da177e4
LT
168static u8 adpt_read_blink_led(adpt_hba* host)
169{
172c122d 170 if (host->FwDebugBLEDflag_P) {
1da177e4
LT
171 if( readb(host->FwDebugBLEDflag_P) == 0xbc ){
172 return readb(host->FwDebugBLEDvalue_P);
173 }
174 }
175 return 0;
176}
177
178/*============================================================================
179 * Scsi host template interface functions
180 *============================================================================
181 */
182
f50abb9b 183#ifdef MODULE
1da177e4
LT
184static struct pci_device_id dptids[] = {
185 { PCI_DPT_VENDOR_ID, PCI_DPT_DEVICE_ID, PCI_ANY_ID, PCI_ANY_ID,},
186 { PCI_DPT_VENDOR_ID, PCI_DPT_RAPTOR_DEVICE_ID, PCI_ANY_ID, PCI_ANY_ID,},
187 { 0, }
188};
f50abb9b
SM
189#endif
190
1da177e4
LT
191MODULE_DEVICE_TABLE(pci,dptids);
192
24601bbc 193static int adpt_detect(struct scsi_host_template* sht)
1da177e4
LT
194{
195 struct pci_dev *pDev = NULL;
229bab6b
DC
196 adpt_hba *pHba;
197 adpt_hba *next;
1da177e4 198
1da177e4
LT
199 PINFO("Detecting Adaptec I2O RAID controllers...\n");
200
201 /* search for all Adatpec I2O RAID cards */
a07f3537 202 while ((pDev = pci_get_device( PCI_DPT_VENDOR_ID, PCI_ANY_ID, pDev))) {
1da177e4
LT
203 if(pDev->device == PCI_DPT_DEVICE_ID ||
204 pDev->device == PCI_DPT_RAPTOR_DEVICE_ID){
24601bbc 205 if(adpt_install_hba(sht, pDev) ){
1da177e4
LT
206 PERROR("Could not Init an I2O RAID device\n");
207 PERROR("Will not try to detect others.\n");
208 return hba_count-1;
209 }
a07f3537 210 pci_dev_get(pDev);
1da177e4
LT
211 }
212 }
213
214 /* In INIT state, Activate IOPs */
229bab6b
DC
215 for (pHba = hba_chain; pHba; pHba = next) {
216 next = pHba->next;
1da177e4
LT
217 // Activate does get status , init outbound, and get hrt
218 if (adpt_i2o_activate_hba(pHba) < 0) {
219 adpt_i2o_delete_hba(pHba);
220 }
221 }
222
223
224 /* Active IOPs in HOLD state */
225
226rebuild_sys_tab:
227 if (hba_chain == NULL)
228 return 0;
229
230 /*
231 * If build_sys_table fails, we kill everything and bail
232 * as we can't init the IOPs w/o a system table
233 */
234 if (adpt_i2o_build_sys_table() < 0) {
235 adpt_i2o_sys_shutdown();
236 return 0;
237 }
238
239 PDEBUG("HBA's in HOLD state\n");
240
241 /* If IOP don't get online, we need to rebuild the System table */
242 for (pHba = hba_chain; pHba; pHba = pHba->next) {
243 if (adpt_i2o_online_hba(pHba) < 0) {
244 adpt_i2o_delete_hba(pHba);
245 goto rebuild_sys_tab;
246 }
247 }
248
249 /* Active IOPs now in OPERATIONAL state */
250 PDEBUG("HBA's in OPERATIONAL state\n");
251
252 printk("dpti: If you have a lot of devices this could take a few minutes.\n");
229bab6b
DC
253 for (pHba = hba_chain; pHba; pHba = next) {
254 next = pHba->next;
1da177e4
LT
255 printk(KERN_INFO"%s: Reading the hardware resource table.\n", pHba->name);
256 if (adpt_i2o_lct_get(pHba) < 0){
257 adpt_i2o_delete_hba(pHba);
258 continue;
259 }
260
261 if (adpt_i2o_parse_lct(pHba) < 0){
262 adpt_i2o_delete_hba(pHba);
263 continue;
264 }
265 adpt_inquiry(pHba);
266 }
267
1ed43910
MS
268 adpt_sysfs_class = class_create(THIS_MODULE, "dpt_i2o");
269 if (IS_ERR(adpt_sysfs_class)) {
270 printk(KERN_WARNING"dpti: unable to create dpt_i2o class\n");
271 adpt_sysfs_class = NULL;
272 }
273
229bab6b
DC
274 for (pHba = hba_chain; pHba; pHba = next) {
275 next = pHba->next;
c864cb14 276 if (adpt_scsi_host_alloc(pHba, sht) < 0){
1da177e4
LT
277 adpt_i2o_delete_hba(pHba);
278 continue;
279 }
280 pHba->initialized = TRUE;
281 pHba->state &= ~DPTI_STATE_RESET;
1ed43910 282 if (adpt_sysfs_class) {
d73a1a67 283 struct device *dev = device_create(adpt_sysfs_class,
9def0b97 284 NULL, MKDEV(DPTI_I2O_MAJOR, pHba->unit), NULL,
1ed43910
MS
285 "dpti%d", pHba->unit);
286 if (IS_ERR(dev)) {
287 printk(KERN_WARNING"dpti%d: unable to "
288 "create device in dpt_i2o class\n",
289 pHba->unit);
290 }
291 }
1da177e4
LT
292 }
293
294 // Register our control device node
295 // nodes will need to be created in /dev to access this
296 // the nodes can not be created from within the driver
297 if (hba_count && register_chrdev(DPTI_I2O_MAJOR, DPT_DRIVER, &adpt_fops)) {
24601bbc 298 adpt_i2o_sys_shutdown();
1da177e4
LT
299 return 0;
300 }
301 return hba_count;
302}
303
304
24601bbc
AM
305/*
306 * scsi_unregister will be called AFTER we return.
307 */
308static int adpt_release(struct Scsi_Host *host)
1da177e4 309{
24601bbc 310 adpt_hba* pHba = (adpt_hba*) host->hostdata[0];
1da177e4
LT
311// adpt_i2o_quiesce_hba(pHba);
312 adpt_i2o_delete_hba(pHba);
24601bbc 313 scsi_unregister(host);
1da177e4
LT
314 return 0;
315}
316
317
318static void adpt_inquiry(adpt_hba* pHba)
319{
62ac5aed 320 u32 msg[17];
1da177e4
LT
321 u32 *mptr;
322 u32 *lenptr;
323 int direction;
324 int scsidir;
325 u32 len;
326 u32 reqlen;
327 u8* buf;
67af2b06 328 dma_addr_t addr;
1da177e4
LT
329 u8 scb[16];
330 s32 rcode;
331
332 memset(msg, 0, sizeof(msg));
67af2b06 333 buf = dma_alloc_coherent(&pHba->pDev->dev, 80, &addr, GFP_KERNEL);
1da177e4
LT
334 if(!buf){
335 printk(KERN_ERR"%s: Could not allocate buffer\n",pHba->name);
336 return;
337 }
338 memset((void*)buf, 0, 36);
339
340 len = 36;
341 direction = 0x00000000;
342 scsidir =0x40000000; // DATA IN (iop<--dev)
343
62ac5aed
MS
344 if (dpt_dma64(pHba))
345 reqlen = 17; // SINGLE SGE, 64 bit
346 else
347 reqlen = 14; // SINGLE SGE, 32 bit
1da177e4
LT
348 /* Stick the headers on */
349 msg[0] = reqlen<<16 | SGL_OFFSET_12;
350 msg[1] = (0xff<<24|HOST_TID<<12|ADAPTER_TID);
351 msg[2] = 0;
352 msg[3] = 0;
353 // Adaptec/DPT Private stuff
354 msg[4] = I2O_CMD_SCSI_EXEC|DPT_ORGANIZATION_ID<<16;
355 msg[5] = ADAPTER_TID | 1<<16 /* Interpret*/;
356 /* Direction, disconnect ok | sense data | simple queue , CDBLen */
357 // I2O_SCB_FLAG_ENABLE_DISCONNECT |
358 // I2O_SCB_FLAG_SIMPLE_QUEUE_TAG |
359 // I2O_SCB_FLAG_SENSE_DATA_IN_MESSAGE;
360 msg[6] = scsidir|0x20a00000| 6 /* cmd len*/;
361
362 mptr=msg+7;
363
364 memset(scb, 0, sizeof(scb));
365 // Write SCSI command into the message - always 16 byte block
366 scb[0] = INQUIRY;
367 scb[1] = 0;
368 scb[2] = 0;
369 scb[3] = 0;
370 scb[4] = 36;
371 scb[5] = 0;
372 // Don't care about the rest of scb
373
374 memcpy(mptr, scb, sizeof(scb));
375 mptr+=4;
376 lenptr=mptr++; /* Remember me - fill in when we know */
377
378 /* Now fill in the SGList and command */
379 *lenptr = len;
62ac5aed
MS
380 if (dpt_dma64(pHba)) {
381 *mptr++ = (0x7C<<24)+(2<<16)+0x02; /* Enable 64 bit */
382 *mptr++ = 1 << PAGE_SHIFT;
383 *mptr++ = 0xD0000000|direction|len;
384 *mptr++ = dma_low(addr);
385 *mptr++ = dma_high(addr);
386 } else {
387 *mptr++ = 0xD0000000|direction|len;
388 *mptr++ = addr;
389 }
1da177e4
LT
390
391 // Send it on it's way
392 rcode = adpt_i2o_post_wait(pHba, msg, reqlen<<2, 120);
393 if (rcode != 0) {
394 sprintf(pHba->detail, "Adaptec I2O RAID");
395 printk(KERN_INFO "%s: Inquiry Error (%d)\n",pHba->name,rcode);
396 if (rcode != -ETIME && rcode != -EINTR)
67af2b06 397 dma_free_coherent(&pHba->pDev->dev, 80, buf, addr);
1da177e4
LT
398 } else {
399 memset(pHba->detail, 0, sizeof(pHba->detail));
400 memcpy(&(pHba->detail), "Vendor: Adaptec ", 16);
401 memcpy(&(pHba->detail[16]), " Model: ", 8);
402 memcpy(&(pHba->detail[24]), (u8*) &buf[16], 16);
403 memcpy(&(pHba->detail[40]), " FW: ", 4);
404 memcpy(&(pHba->detail[44]), (u8*) &buf[32], 4);
405 pHba->detail[48] = '\0'; /* precautionary */
67af2b06 406 dma_free_coherent(&pHba->pDev->dev, 80, buf, addr);
1da177e4
LT
407 }
408 adpt_i2o_status_get(pHba);
409 return ;
410}
411
412
413static int adpt_slave_configure(struct scsi_device * device)
414{
415 struct Scsi_Host *host = device->host;
416 adpt_hba* pHba;
417
418 pHba = (adpt_hba *) host->hostdata[0];
419
420 if (host->can_queue && device->tagged_supported) {
db5ed4df 421 scsi_change_queue_depth(device,
1da177e4 422 host->can_queue - 1);
1da177e4
LT
423 }
424 return 0;
425}
426
f281233d 427static int adpt_queue_lck(struct scsi_cmnd * cmd, void (*done) (struct scsi_cmnd *))
1da177e4
LT
428{
429 adpt_hba* pHba = NULL;
430 struct adpt_device* pDev = NULL; /* dpt per device information */
1da177e4
LT
431
432 cmd->scsi_done = done;
433 /*
434 * SCSI REQUEST_SENSE commands will be executed automatically by the
435 * Host Adapter for any errors, so they should not be executed
436 * explicitly unless the Sense Data is zero indicating that no error
437 * occurred.
438 */
439
440 if ((cmd->cmnd[0] == REQUEST_SENSE) && (cmd->sense_buffer[0] != 0)) {
441 cmd->result = (DID_OK << 16);
442 cmd->scsi_done(cmd);
443 return 0;
444 }
445
446 pHba = (adpt_hba*)cmd->device->host->hostdata[0];
447 if (!pHba) {
448 return FAILED;
449 }
450
451 rmb();
63d80c49
HR
452 if ((pHba->state) & DPTI_STATE_RESET)
453 return SCSI_MLQUEUE_HOST_BUSY;
1da177e4 454
1da177e4
LT
455 // TODO if the cmd->device if offline then I may need to issue a bus rescan
456 // followed by a get_lct to see if the device is there anymore
457 if((pDev = (struct adpt_device*) (cmd->device->hostdata)) == NULL) {
458 /*
459 * First command request for this device. Set up a pointer
460 * to the device structure. This should be a TEST_UNIT_READY
461 * command from scan_scsis_single.
462 */
9cb78c16 463 if ((pDev = adpt_find_device(pHba, (u32)cmd->device->channel, (u32)cmd->device->id, cmd->device->lun)) == NULL) {
1da177e4
LT
464 // TODO: if any luns are at this bus, scsi id then fake a TEST_UNIT_READY and INQUIRY response
465 // with type 7F (for all luns less than the max for this bus,id) so the lun scan will continue.
466 cmd->result = (DID_NO_CONNECT << 16);
467 cmd->scsi_done(cmd);
468 return 0;
469 }
470 cmd->device->hostdata = pDev;
471 }
472 pDev->pScsi_dev = cmd->device;
473
474 /*
475 * If we are being called from when the device is being reset,
476 * delay processing of the command until later.
477 */
478 if (pDev->state & DPTI_DEV_RESET ) {
479 return FAILED;
480 }
481 return adpt_scsi_to_i2o(pHba, cmd, pDev);
482}
483
f281233d
JG
484static DEF_SCSI_QCMD(adpt_queue)
485
1da177e4
LT
486static int adpt_bios_param(struct scsi_device *sdev, struct block_device *dev,
487 sector_t capacity, int geom[])
488{
489 int heads=-1;
490 int sectors=-1;
491 int cylinders=-1;
492
493 // *** First lets set the default geometry ****
494
495 // If the capacity is less than ox2000
496 if (capacity < 0x2000 ) { // floppy
497 heads = 18;
498 sectors = 2;
499 }
500 // else if between 0x2000 and 0x20000
501 else if (capacity < 0x20000) {
502 heads = 64;
503 sectors = 32;
504 }
505 // else if between 0x20000 and 0x40000
506 else if (capacity < 0x40000) {
507 heads = 65;
508 sectors = 63;
509 }
510 // else if between 0x4000 and 0x80000
511 else if (capacity < 0x80000) {
512 heads = 128;
513 sectors = 63;
514 }
515 // else if greater than 0x80000
516 else {
517 heads = 255;
518 sectors = 63;
519 }
520 cylinders = sector_div(capacity, heads * sectors);
521
522 // Special case if CDROM
523 if(sdev->type == 5) { // CDROM
524 heads = 252;
525 sectors = 63;
526 cylinders = 1111;
527 }
528
529 geom[0] = heads;
530 geom[1] = sectors;
531 geom[2] = cylinders;
532
533 PDEBUG("adpt_bios_param: exit\n");
534 return 0;
535}
536
537
538static const char *adpt_info(struct Scsi_Host *host)
539{
540 adpt_hba* pHba;
541
542 pHba = (adpt_hba *) host->hostdata[0];
543 return (char *) (pHba->detail);
544}
545
ff98f7ce 546static int adpt_show_info(struct seq_file *m, struct Scsi_Host *host)
1da177e4
LT
547{
548 struct adpt_device* d;
549 int id;
550 int chan;
1da177e4
LT
551 adpt_hba* pHba;
552 int unit;
553
1da177e4 554 // Find HBA (host bus adapter) we are looking for
0b950672 555 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
556 for (pHba = hba_chain; pHba; pHba = pHba->next) {
557 if (pHba->host == host) {
558 break; /* found adapter */
559 }
560 }
0b950672 561 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
562 if (pHba == NULL) {
563 return 0;
564 }
565 host = pHba->host;
566
ff98f7ce
AV
567 seq_printf(m, "Adaptec I2O RAID Driver Version: %s\n\n", DPT_I2O_VERSION);
568 seq_printf(m, "%s\n", pHba->detail);
569 seq_printf(m, "SCSI Host=scsi%d Control Node=/dev/%s irq=%d\n",
1da177e4 570 pHba->host->host_no, pHba->name, host->irq);
ff98f7ce 571 seq_printf(m, "\tpost fifo size = %d\n\treply fifo size = %d\n\tsg table size = %d\n\n",
1da177e4
LT
572 host->can_queue, (int) pHba->reply_fifo_size , host->sg_tablesize);
573
91c40f24 574 seq_puts(m, "Devices:\n");
1da177e4
LT
575 for(chan = 0; chan < MAX_CHANNEL; chan++) {
576 for(id = 0; id < MAX_ID; id++) {
577 d = pHba->channel[chan].device[id];
ff98f7ce
AV
578 while(d) {
579 seq_printf(m,"\t%-24.24s", d->pScsi_dev->vendor);
580 seq_printf(m," Rev: %-8.8s\n", d->pScsi_dev->rev);
1da177e4
LT
581
582 unit = d->pI2o_dev->lct_data.tid;
9cb78c16
HR
583 seq_printf(m, "\tTID=%d, (Channel=%d, Target=%d, Lun=%llu) (%s)\n\n",
584 unit, (int)d->scsi_channel, (int)d->scsi_id, d->scsi_lun,
1da177e4 585 scsi_device_online(d->pScsi_dev)? "online":"offline");
1da177e4
LT
586 d = d->next_lun;
587 }
588 }
589 }
ff98f7ce 590 return 0;
1da177e4
LT
591}
592
62ac5aed
MS
593/*
594 * Turn a struct scsi_cmnd * into a unique 32 bit 'context'.
595 */
596static u32 adpt_cmd_to_context(struct scsi_cmnd *cmd)
597{
598 return (u32)cmd->serial_number;
599}
600
601/*
602 * Go from a u32 'context' to a struct scsi_cmnd * .
603 * This could probably be made more efficient.
604 */
605static struct scsi_cmnd *
606 adpt_cmd_from_context(adpt_hba * pHba, u32 context)
607{
608 struct scsi_cmnd * cmd;
609 struct scsi_device * d;
610
611 if (context == 0)
612 return NULL;
613
614 spin_unlock(pHba->host->host_lock);
615 shost_for_each_device(d, pHba->host) {
616 unsigned long flags;
617 spin_lock_irqsave(&d->list_lock, flags);
618 list_for_each_entry(cmd, &d->cmd_list, list) {
619 if (((u32)cmd->serial_number == context)) {
620 spin_unlock_irqrestore(&d->list_lock, flags);
621 scsi_device_put(d);
622 spin_lock(pHba->host->host_lock);
623 return cmd;
624 }
625 }
626 spin_unlock_irqrestore(&d->list_lock, flags);
627 }
628 spin_lock(pHba->host->host_lock);
629
630 return NULL;
631}
632
633/*
634 * Turn a pointer to ioctl reply data into an u32 'context'
635 */
636static u32 adpt_ioctl_to_context(adpt_hba * pHba, void *reply)
637{
638#if BITS_PER_LONG == 32
639 return (u32)(unsigned long)reply;
640#else
641 ulong flags = 0;
642 u32 nr, i;
643
644 spin_lock_irqsave(pHba->host->host_lock, flags);
645 nr = ARRAY_SIZE(pHba->ioctl_reply_context);
646 for (i = 0; i < nr; i++) {
647 if (pHba->ioctl_reply_context[i] == NULL) {
648 pHba->ioctl_reply_context[i] = reply;
649 break;
650 }
651 }
652 spin_unlock_irqrestore(pHba->host->host_lock, flags);
653 if (i >= nr) {
62ac5aed
MS
654 printk(KERN_WARNING"%s: Too many outstanding "
655 "ioctl commands\n", pHba->name);
656 return (u32)-1;
657 }
658
659 return i;
660#endif
661}
662
663/*
664 * Go from an u32 'context' to a pointer to ioctl reply data.
665 */
666static void *adpt_ioctl_from_context(adpt_hba *pHba, u32 context)
667{
668#if BITS_PER_LONG == 32
669 return (void *)(unsigned long)context;
670#else
671 void *p = pHba->ioctl_reply_context[context];
672 pHba->ioctl_reply_context[context] = NULL;
673
674 return p;
675#endif
676}
1da177e4
LT
677
678/*===========================================================================
679 * Error Handling routines
680 *===========================================================================
681 */
682
683static int adpt_abort(struct scsi_cmnd * cmd)
684{
685 adpt_hba* pHba = NULL; /* host bus adapter structure */
686 struct adpt_device* dptdevice; /* dpt per device information */
687 u32 msg[5];
688 int rcode;
689
690 if(cmd->serial_number == 0){
691 return FAILED;
692 }
693 pHba = (adpt_hba*) cmd->device->host->hostdata[0];
5cd049a5 694 printk(KERN_INFO"%s: Trying to Abort\n",pHba->name);
1da177e4
LT
695 if ((dptdevice = (void*) (cmd->device->hostdata)) == NULL) {
696 printk(KERN_ERR "%s: Unable to abort: No device in cmnd\n",pHba->name);
697 return FAILED;
698 }
699
700 memset(msg, 0, sizeof(msg));
701 msg[0] = FIVE_WORD_MSG_SIZE|SGL_OFFSET_0;
702 msg[1] = I2O_CMD_SCSI_ABORT<<24|HOST_TID<<12|dptdevice->tid;
703 msg[2] = 0;
704 msg[3]= 0;
62ac5aed 705 msg[4] = adpt_cmd_to_context(cmd);
e5508c13
SM
706 if (pHba->host)
707 spin_lock_irq(pHba->host->host_lock);
708 rcode = adpt_i2o_post_wait(pHba, msg, sizeof(msg), FOREVER);
709 if (pHba->host)
710 spin_unlock_irq(pHba->host->host_lock);
711 if (rcode != 0) {
1da177e4
LT
712 if(rcode == -EOPNOTSUPP ){
713 printk(KERN_INFO"%s: Abort cmd not supported\n",pHba->name);
714 return FAILED;
715 }
5cd049a5 716 printk(KERN_INFO"%s: Abort failed.\n",pHba->name);
1da177e4
LT
717 return FAILED;
718 }
5cd049a5 719 printk(KERN_INFO"%s: Abort complete.\n",pHba->name);
1da177e4
LT
720 return SUCCESS;
721}
722
723
724#define I2O_DEVICE_RESET 0x27
725// This is the same for BLK and SCSI devices
726// NOTE this is wrong in the i2o.h definitions
727// This is not currently supported by our adapter but we issue it anyway
728static int adpt_device_reset(struct scsi_cmnd* cmd)
729{
730 adpt_hba* pHba;
731 u32 msg[4];
732 u32 rcode;
733 int old_state;
1c2fb3f3 734 struct adpt_device* d = cmd->device->hostdata;
1da177e4
LT
735
736 pHba = (void*) cmd->device->host->hostdata[0];
737 printk(KERN_INFO"%s: Trying to reset device\n",pHba->name);
738 if (!d) {
739 printk(KERN_INFO"%s: Reset Device: Device Not found\n",pHba->name);
740 return FAILED;
741 }
742 memset(msg, 0, sizeof(msg));
743 msg[0] = FOUR_WORD_MSG_SIZE|SGL_OFFSET_0;
744 msg[1] = (I2O_DEVICE_RESET<<24|HOST_TID<<12|d->tid);
745 msg[2] = 0;
746 msg[3] = 0;
747
e5508c13
SM
748 if (pHba->host)
749 spin_lock_irq(pHba->host->host_lock);
1da177e4
LT
750 old_state = d->state;
751 d->state |= DPTI_DEV_RESET;
e5508c13
SM
752 rcode = adpt_i2o_post_wait(pHba, msg,sizeof(msg), FOREVER);
753 d->state = old_state;
754 if (pHba->host)
755 spin_unlock_irq(pHba->host->host_lock);
756 if (rcode != 0) {
1da177e4
LT
757 if(rcode == -EOPNOTSUPP ){
758 printk(KERN_INFO"%s: Device reset not supported\n",pHba->name);
759 return FAILED;
760 }
761 printk(KERN_INFO"%s: Device reset failed\n",pHba->name);
762 return FAILED;
763 } else {
1da177e4
LT
764 printk(KERN_INFO"%s: Device reset successful\n",pHba->name);
765 return SUCCESS;
766 }
767}
768
769
770#define I2O_HBA_BUS_RESET 0x87
771// This version of bus reset is called by the eh_error handler
772static int adpt_bus_reset(struct scsi_cmnd* cmd)
773{
774 adpt_hba* pHba;
775 u32 msg[4];
e5508c13 776 u32 rcode;
1da177e4
LT
777
778 pHba = (adpt_hba*)cmd->device->host->hostdata[0];
779 memset(msg, 0, sizeof(msg));
780 printk(KERN_WARNING"%s: Bus reset: SCSI Bus %d: tid: %d\n",pHba->name, cmd->device->channel,pHba->channel[cmd->device->channel].tid );
781 msg[0] = FOUR_WORD_MSG_SIZE|SGL_OFFSET_0;
782 msg[1] = (I2O_HBA_BUS_RESET<<24|HOST_TID<<12|pHba->channel[cmd->device->channel].tid);
783 msg[2] = 0;
784 msg[3] = 0;
e5508c13
SM
785 if (pHba->host)
786 spin_lock_irq(pHba->host->host_lock);
787 rcode = adpt_i2o_post_wait(pHba, msg,sizeof(msg), FOREVER);
788 if (pHba->host)
789 spin_unlock_irq(pHba->host->host_lock);
790 if (rcode != 0) {
1da177e4
LT
791 printk(KERN_WARNING"%s: Bus reset failed.\n",pHba->name);
792 return FAILED;
793 } else {
794 printk(KERN_WARNING"%s: Bus reset success.\n",pHba->name);
795 return SUCCESS;
796 }
797}
798
799// This version of reset is called by the eh_error_handler
df0ae249 800static int __adpt_reset(struct scsi_cmnd* cmd)
1da177e4
LT
801{
802 adpt_hba* pHba;
803 int rcode;
804 pHba = (adpt_hba*)cmd->device->host->hostdata[0];
805 printk(KERN_WARNING"%s: Hba Reset: scsi id %d: tid: %d\n",pHba->name,cmd->device->channel,pHba->channel[cmd->device->channel].tid );
806 rcode = adpt_hba_reset(pHba);
807 if(rcode == 0){
808 printk(KERN_WARNING"%s: HBA reset complete\n",pHba->name);
809 return SUCCESS;
810 } else {
811 printk(KERN_WARNING"%s: HBA reset failed (%x)\n",pHba->name, rcode);
812 return FAILED;
813 }
814}
815
df0ae249
JG
816static int adpt_reset(struct scsi_cmnd* cmd)
817{
818 int rc;
819
820 spin_lock_irq(cmd->device->host->host_lock);
821 rc = __adpt_reset(cmd);
822 spin_unlock_irq(cmd->device->host->host_lock);
823
824 return rc;
825}
826
1da177e4
LT
827// This version of reset is called by the ioctls and indirectly from eh_error_handler via adpt_reset
828static int adpt_hba_reset(adpt_hba* pHba)
829{
830 int rcode;
831
832 pHba->state |= DPTI_STATE_RESET;
833
834 // Activate does get status , init outbound, and get hrt
835 if ((rcode=adpt_i2o_activate_hba(pHba)) < 0) {
836 printk(KERN_ERR "%s: Could not activate\n", pHba->name);
837 adpt_i2o_delete_hba(pHba);
838 return rcode;
839 }
840
841 if ((rcode=adpt_i2o_build_sys_table()) < 0) {
842 adpt_i2o_delete_hba(pHba);
843 return rcode;
844 }
845 PDEBUG("%s: in HOLD state\n",pHba->name);
846
847 if ((rcode=adpt_i2o_online_hba(pHba)) < 0) {
848 adpt_i2o_delete_hba(pHba);
849 return rcode;
850 }
851 PDEBUG("%s: in OPERATIONAL state\n",pHba->name);
852
853 if ((rcode=adpt_i2o_lct_get(pHba)) < 0){
854 adpt_i2o_delete_hba(pHba);
855 return rcode;
856 }
857
858 if ((rcode=adpt_i2o_reparse_lct(pHba)) < 0){
859 adpt_i2o_delete_hba(pHba);
860 return rcode;
861 }
862 pHba->state &= ~DPTI_STATE_RESET;
863
864 adpt_fail_posted_scbs(pHba);
865 return 0; /* return success */
866}
867
868/*===========================================================================
869 *
870 *===========================================================================
871 */
872
873
874static void adpt_i2o_sys_shutdown(void)
875{
876 adpt_hba *pHba, *pNext;
458af543 877 struct adpt_i2o_post_wait_data *p1, *old;
1da177e4
LT
878
879 printk(KERN_INFO"Shutting down Adaptec I2O controllers.\n");
880 printk(KERN_INFO" This could take a few minutes if there are many devices attached\n");
881 /* Delete all IOPs from the controller chain */
882 /* They should have already been released by the
883 * scsi-core
884 */
885 for (pHba = hba_chain; pHba; pHba = pNext) {
886 pNext = pHba->next;
887 adpt_i2o_delete_hba(pHba);
888 }
889
890 /* Remove any timedout entries from the wait queue. */
1da177e4
LT
891// spin_lock_irqsave(&adpt_post_wait_lock, flags);
892 /* Nothing should be outstanding at this point so just
893 * free them
894 */
458af543
AB
895 for(p1 = adpt_post_wait_queue; p1;) {
896 old = p1;
897 p1 = p1->next;
898 kfree(old);
1da177e4
LT
899 }
900// spin_unlock_irqrestore(&adpt_post_wait_lock, flags);
901 adpt_post_wait_queue = NULL;
902
903 printk(KERN_INFO "Adaptec I2O controllers down.\n");
904}
905
24601bbc 906static int adpt_install_hba(struct scsi_host_template* sht, struct pci_dev* pDev)
1da177e4
LT
907{
908
909 adpt_hba* pHba = NULL;
910 adpt_hba* p = NULL;
911 ulong base_addr0_phys = 0;
912 ulong base_addr1_phys = 0;
913 u32 hba_map0_area_size = 0;
914 u32 hba_map1_area_size = 0;
915 void __iomem *base_addr_virt = NULL;
916 void __iomem *msg_addr_virt = NULL;
62ac5aed 917 int dma64 = 0;
1da177e4
LT
918
919 int raptorFlag = FALSE;
1da177e4
LT
920
921 if(pci_enable_device(pDev)) {
922 return -EINVAL;
923 }
9638d89a
SM
924
925 if (pci_request_regions(pDev, "dpt_i2o")) {
926 PERROR("dpti: adpt_config_hba: pci request region failed\n");
927 return -EINVAL;
928 }
929
1da177e4 930 pci_set_master(pDev);
62ac5aed
MS
931
932 /*
933 * See if we should enable dma64 mode.
934 */
935 if (sizeof(dma_addr_t) > 4 &&
6a35528a 936 pci_set_dma_mask(pDev, DMA_BIT_MASK(64)) == 0) {
284901a9 937 if (dma_get_required_mask(&pDev->dev) > DMA_BIT_MASK(32))
62ac5aed
MS
938 dma64 = 1;
939 }
284901a9 940 if (!dma64 && pci_set_dma_mask(pDev, DMA_BIT_MASK(32)) != 0)
1da177e4
LT
941 return -EINVAL;
942
67af2b06 943 /* adapter only supports message blocks below 4GB */
284901a9 944 pci_set_consistent_dma_mask(pDev, DMA_BIT_MASK(32));
67af2b06 945
1da177e4
LT
946 base_addr0_phys = pci_resource_start(pDev,0);
947 hba_map0_area_size = pci_resource_len(pDev,0);
948
949 // Check if standard PCI card or single BAR Raptor
950 if(pDev->device == PCI_DPT_DEVICE_ID){
951 if(pDev->subsystem_device >=0xc032 && pDev->subsystem_device <= 0xc03b){
952 // Raptor card with this device id needs 4M
953 hba_map0_area_size = 0x400000;
954 } else { // Not Raptor - it is a PCI card
955 if(hba_map0_area_size > 0x100000 ){
956 hba_map0_area_size = 0x100000;
957 }
958 }
959 } else {// Raptor split BAR config
960 // Use BAR1 in this configuration
961 base_addr1_phys = pci_resource_start(pDev,1);
962 hba_map1_area_size = pci_resource_len(pDev,1);
963 raptorFlag = TRUE;
964 }
965
62ac5aed
MS
966#if BITS_PER_LONG == 64
967 /*
968 * The original Adaptec 64 bit driver has this comment here:
969 * "x86_64 machines need more optimal mappings"
970 *
971 * I assume some HBAs report ridiculously large mappings
972 * and we need to limit them on platforms with IOMMUs.
973 */
974 if (raptorFlag == TRUE) {
975 if (hba_map0_area_size > 128)
976 hba_map0_area_size = 128;
977 if (hba_map1_area_size > 524288)
978 hba_map1_area_size = 524288;
979 } else {
980 if (hba_map0_area_size > 524288)
981 hba_map0_area_size = 524288;
982 }
983#endif
984
1da177e4
LT
985 base_addr_virt = ioremap(base_addr0_phys,hba_map0_area_size);
986 if (!base_addr_virt) {
9c472dd9 987 pci_release_regions(pDev);
1da177e4
LT
988 PERROR("dpti: adpt_config_hba: io remap failed\n");
989 return -EINVAL;
990 }
991
992 if(raptorFlag == TRUE) {
993 msg_addr_virt = ioremap(base_addr1_phys, hba_map1_area_size );
994 if (!msg_addr_virt) {
995 PERROR("dpti: adpt_config_hba: io remap failed on BAR1\n");
996 iounmap(base_addr_virt);
9c472dd9 997 pci_release_regions(pDev);
1da177e4
LT
998 return -EINVAL;
999 }
1000 } else {
1001 msg_addr_virt = base_addr_virt;
1002 }
1003
1004 // Allocate and zero the data structure
bbfbbbc1
MK
1005 pHba = kzalloc(sizeof(adpt_hba), GFP_KERNEL);
1006 if (!pHba) {
1007 if (msg_addr_virt != base_addr_virt)
1da177e4 1008 iounmap(msg_addr_virt);
1da177e4 1009 iounmap(base_addr_virt);
9c472dd9 1010 pci_release_regions(pDev);
1da177e4
LT
1011 return -ENOMEM;
1012 }
1da177e4 1013
0b950672 1014 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
1015
1016 if(hba_chain != NULL){
1017 for(p = hba_chain; p->next; p = p->next);
1018 p->next = pHba;
1019 } else {
1020 hba_chain = pHba;
1021 }
1022 pHba->next = NULL;
1023 pHba->unit = hba_count;
23a2bc22 1024 sprintf(pHba->name, "dpti%d", hba_count);
1da177e4
LT
1025 hba_count++;
1026
0b950672 1027 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
1028
1029 pHba->pDev = pDev;
1030 pHba->base_addr_phys = base_addr0_phys;
1031
1032 // Set up the Virtual Base Address of the I2O Device
1033 pHba->base_addr_virt = base_addr_virt;
1034 pHba->msg_addr_virt = msg_addr_virt;
1035 pHba->irq_mask = base_addr_virt+0x30;
1036 pHba->post_port = base_addr_virt+0x40;
1037 pHba->reply_port = base_addr_virt+0x44;
1038
1039 pHba->hrt = NULL;
1040 pHba->lct = NULL;
1041 pHba->lct_size = 0;
1042 pHba->status_block = NULL;
1043 pHba->post_count = 0;
1044 pHba->state = DPTI_STATE_RESET;
1045 pHba->pDev = pDev;
1046 pHba->devices = NULL;
62ac5aed 1047 pHba->dma64 = dma64;
1da177e4
LT
1048
1049 // Initializing the spinlocks
1050 spin_lock_init(&pHba->state_lock);
1051 spin_lock_init(&adpt_post_wait_lock);
1052
1053 if(raptorFlag == 0){
62ac5aed
MS
1054 printk(KERN_INFO "Adaptec I2O RAID controller"
1055 " %d at %p size=%x irq=%d%s\n",
1056 hba_count-1, base_addr_virt,
1057 hba_map0_area_size, pDev->irq,
1058 dma64 ? " (64-bit DMA)" : "");
1da177e4 1059 } else {
62ac5aed
MS
1060 printk(KERN_INFO"Adaptec I2O RAID controller %d irq=%d%s\n",
1061 hba_count-1, pDev->irq,
1062 dma64 ? " (64-bit DMA)" : "");
1da177e4
LT
1063 printk(KERN_INFO" BAR0 %p - size= %x\n",base_addr_virt,hba_map0_area_size);
1064 printk(KERN_INFO" BAR1 %p - size= %x\n",msg_addr_virt,hba_map1_area_size);
1065 }
1066
1d6f359a 1067 if (request_irq (pDev->irq, adpt_isr, IRQF_SHARED, pHba->name, pHba)) {
1da177e4
LT
1068 printk(KERN_ERR"%s: Couldn't register IRQ %d\n", pHba->name, pDev->irq);
1069 adpt_i2o_delete_hba(pHba);
1070 return -EINVAL;
1071 }
1072
1073 return 0;
1074}
1075
1076
1077static void adpt_i2o_delete_hba(adpt_hba* pHba)
1078{
1079 adpt_hba* p1;
1080 adpt_hba* p2;
1081 struct i2o_device* d;
1082 struct i2o_device* next;
1083 int i;
1084 int j;
1085 struct adpt_device* pDev;
1086 struct adpt_device* pNext;
1087
1088
0b950672 1089 mutex_lock(&adpt_configuration_lock);
24601bbc
AM
1090 // scsi_unregister calls our adpt_release which
1091 // does a quiese
1da177e4
LT
1092 if(pHba->host){
1093 free_irq(pHba->host->irq, pHba);
1094 }
1da177e4
LT
1095 p2 = NULL;
1096 for( p1 = hba_chain; p1; p2 = p1,p1=p1->next){
1097 if(p1 == pHba) {
1098 if(p2) {
1099 p2->next = p1->next;
1100 } else {
1101 hba_chain = p1->next;
1102 }
1103 break;
1104 }
1105 }
1106
1107 hba_count--;
0b950672 1108 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
1109
1110 iounmap(pHba->base_addr_virt);
9c472dd9 1111 pci_release_regions(pHba->pDev);
1da177e4
LT
1112 if(pHba->msg_addr_virt != pHba->base_addr_virt){
1113 iounmap(pHba->msg_addr_virt);
1114 }
62ac5aed
MS
1115 if(pHba->FwDebugBuffer_P)
1116 iounmap(pHba->FwDebugBuffer_P);
67af2b06
MS
1117 if(pHba->hrt) {
1118 dma_free_coherent(&pHba->pDev->dev,
1119 pHba->hrt->num_entries * pHba->hrt->entry_len << 2,
1120 pHba->hrt, pHba->hrt_pa);
1121 }
1122 if(pHba->lct) {
1123 dma_free_coherent(&pHba->pDev->dev, pHba->lct_size,
1124 pHba->lct, pHba->lct_pa);
1125 }
1126 if(pHba->status_block) {
1127 dma_free_coherent(&pHba->pDev->dev, sizeof(i2o_status_block),
1128 pHba->status_block, pHba->status_block_pa);
1129 }
1130 if(pHba->reply_pool) {
1131 dma_free_coherent(&pHba->pDev->dev,
1132 pHba->reply_fifo_size * REPLY_FRAME_SIZE * 4,
1133 pHba->reply_pool, pHba->reply_pool_pa);
1134 }
1da177e4
LT
1135
1136 for(d = pHba->devices; d ; d = next){
1137 next = d->next;
1138 kfree(d);
1139 }
1140 for(i = 0 ; i < pHba->top_scsi_channel ; i++){
1141 for(j = 0; j < MAX_ID; j++){
1142 if(pHba->channel[i].device[j] != NULL){
1143 for(pDev = pHba->channel[i].device[j]; pDev; pDev = pNext){
1144 pNext = pDev->next_lun;
1145 kfree(pDev);
1146 }
1147 }
1148 }
1149 }
a07f3537 1150 pci_dev_put(pHba->pDev);
1ed43910
MS
1151 if (adpt_sysfs_class)
1152 device_destroy(adpt_sysfs_class,
1153 MKDEV(DPTI_I2O_MAJOR, pHba->unit));
229bab6b 1154 kfree(pHba);
1ed43910 1155
1da177e4
LT
1156 if(hba_count <= 0){
1157 unregister_chrdev(DPTI_I2O_MAJOR, DPT_DRIVER);
1ed43910
MS
1158 if (adpt_sysfs_class) {
1159 class_destroy(adpt_sysfs_class);
1160 adpt_sysfs_class = NULL;
1161 }
1da177e4
LT
1162 }
1163}
1164
9cb78c16 1165static struct adpt_device* adpt_find_device(adpt_hba* pHba, u32 chan, u32 id, u64 lun)
1da177e4
LT
1166{
1167 struct adpt_device* d;
1168
1169 if(chan < 0 || chan >= MAX_CHANNEL)
1170 return NULL;
1171
1172 if( pHba->channel[chan].device == NULL){
1173 printk(KERN_DEBUG"Adaptec I2O RAID: Trying to find device before they are allocated\n");
1174 return NULL;
1175 }
1176
1177 d = pHba->channel[chan].device[id];
1178 if(!d || d->tid == 0) {
1179 return NULL;
1180 }
1181
1182 /* If it is the only lun at that address then this should match*/
1183 if(d->scsi_lun == lun){
1184 return d;
1185 }
1186
1187 /* else we need to look through all the luns */
1188 for(d=d->next_lun ; d ; d = d->next_lun){
1189 if(d->scsi_lun == lun){
1190 return d;
1191 }
1192 }
1193 return NULL;
1194}
1195
1196
1197static int adpt_i2o_post_wait(adpt_hba* pHba, u32* msg, int len, int timeout)
1198{
1199 // I used my own version of the WAIT_QUEUE_HEAD
1200 // to handle some version differences
1201 // When embedded in the kernel this could go back to the vanilla one
1202 ADPT_DECLARE_WAIT_QUEUE_HEAD(adpt_wq_i2o_post);
1203 int status = 0;
1204 ulong flags = 0;
1205 struct adpt_i2o_post_wait_data *p1, *p2;
1206 struct adpt_i2o_post_wait_data *wait_data =
da2907ff 1207 kmalloc(sizeof(struct adpt_i2o_post_wait_data), GFP_ATOMIC);
4452ea50 1208 DECLARE_WAITQUEUE(wait, current);
1da177e4 1209
4452ea50 1210 if (!wait_data)
1da177e4 1211 return -ENOMEM;
4452ea50 1212
1da177e4
LT
1213 /*
1214 * The spin locking is needed to keep anyone from playing
1215 * with the queue pointers and id while we do the same
1216 */
1217 spin_lock_irqsave(&adpt_post_wait_lock, flags);
1218 // TODO we need a MORE unique way of getting ids
1219 // to support async LCT get
1220 wait_data->next = adpt_post_wait_queue;
1221 adpt_post_wait_queue = wait_data;
1222 adpt_post_wait_id++;
1223 adpt_post_wait_id &= 0x7fff;
1224 wait_data->id = adpt_post_wait_id;
1225 spin_unlock_irqrestore(&adpt_post_wait_lock, flags);
1226
1227 wait_data->wq = &adpt_wq_i2o_post;
1228 wait_data->status = -ETIMEDOUT;
1229
4452ea50 1230 add_wait_queue(&adpt_wq_i2o_post, &wait);
1da177e4
LT
1231
1232 msg[2] |= 0x80000000 | ((u32)wait_data->id);
1233 timeout *= HZ;
1234 if((status = adpt_i2o_post_this(pHba, msg, len)) == 0){
1235 set_current_state(TASK_INTERRUPTIBLE);
1236 if(pHba->host)
1237 spin_unlock_irq(pHba->host->host_lock);
1238 if (!timeout)
1239 schedule();
1240 else{
1241 timeout = schedule_timeout(timeout);
1242 if (timeout == 0) {
1243 // I/O issued, but cannot get result in
1244 // specified time. Freeing resorces is
1245 // dangerous.
1246 status = -ETIME;
1247 }
1248 }
1249 if(pHba->host)
1250 spin_lock_irq(pHba->host->host_lock);
1251 }
4452ea50 1252 remove_wait_queue(&adpt_wq_i2o_post, &wait);
1da177e4
LT
1253
1254 if(status == -ETIMEDOUT){
1255 printk(KERN_INFO"dpti%d: POST WAIT TIMEOUT\n",pHba->unit);
1256 // We will have to free the wait_data memory during shutdown
1257 return status;
1258 }
1259
1260 /* Remove the entry from the queue. */
1261 p2 = NULL;
1262 spin_lock_irqsave(&adpt_post_wait_lock, flags);
1263 for(p1 = adpt_post_wait_queue; p1; p2 = p1, p1 = p1->next) {
1264 if(p1 == wait_data) {
1265 if(p1->status == I2O_DETAIL_STATUS_UNSUPPORTED_FUNCTION ) {
1266 status = -EOPNOTSUPP;
1267 }
1268 if(p2) {
1269 p2->next = p1->next;
1270 } else {
1271 adpt_post_wait_queue = p1->next;
1272 }
1273 break;
1274 }
1275 }
1276 spin_unlock_irqrestore(&adpt_post_wait_lock, flags);
1277
1278 kfree(wait_data);
1279
1280 return status;
1281}
1282
1283
1284static s32 adpt_i2o_post_this(adpt_hba* pHba, u32* data, int len)
1285{
1286
1287 u32 m = EMPTY_QUEUE;
1288 u32 __iomem *msg;
1289 ulong timeout = jiffies + 30*HZ;
1290 do {
1291 rmb();
1292 m = readl(pHba->post_port);
1293 if (m != EMPTY_QUEUE) {
1294 break;
1295 }
1296 if(time_after(jiffies,timeout)){
1297 printk(KERN_WARNING"dpti%d: Timeout waiting for message frame!\n", pHba->unit);
1298 return -ETIMEDOUT;
1299 }
a9a3047d 1300 schedule_timeout_uninterruptible(1);
1da177e4
LT
1301 } while(m == EMPTY_QUEUE);
1302
1303 msg = pHba->msg_addr_virt + m;
1304 memcpy_toio(msg, data, len);
1305 wmb();
1306
1307 //post message
1308 writel(m, pHba->post_port);
1309 wmb();
1310
1311 return 0;
1312}
1313
1314
1315static void adpt_i2o_post_wait_complete(u32 context, int status)
1316{
1317 struct adpt_i2o_post_wait_data *p1 = NULL;
1318 /*
1319 * We need to search through the adpt_post_wait
1320 * queue to see if the given message is still
1321 * outstanding. If not, it means that the IOP
1322 * took longer to respond to the message than we
1323 * had allowed and timer has already expired.
1324 * Not much we can do about that except log
1325 * it for debug purposes, increase timeout, and recompile
1326 *
1327 * Lock needed to keep anyone from moving queue pointers
1328 * around while we're looking through them.
1329 */
1330
1331 context &= 0x7fff;
1332
1333 spin_lock(&adpt_post_wait_lock);
1334 for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) {
1335 if(p1->id == context) {
1336 p1->status = status;
1337 spin_unlock(&adpt_post_wait_lock);
1338 wake_up_interruptible(p1->wq);
1339 return;
1340 }
1341 }
1342 spin_unlock(&adpt_post_wait_lock);
1343 // If this happens we lose commands that probably really completed
1344 printk(KERN_DEBUG"dpti: Could Not find task %d in wait queue\n",context);
1345 printk(KERN_DEBUG" Tasks in wait queue:\n");
1346 for(p1 = adpt_post_wait_queue; p1; p1 = p1->next) {
1347 printk(KERN_DEBUG" %d\n",p1->id);
1348 }
1349 return;
1350}
1351
1352static s32 adpt_i2o_reset_hba(adpt_hba* pHba)
1353{
1354 u32 msg[8];
1355 u8* status;
67af2b06 1356 dma_addr_t addr;
1da177e4
LT
1357 u32 m = EMPTY_QUEUE ;
1358 ulong timeout = jiffies + (TMOUT_IOPRESET*HZ);
1359
1360 if(pHba->initialized == FALSE) { // First time reset should be quick
1361 timeout = jiffies + (25*HZ);
1362 } else {
1363 adpt_i2o_quiesce_hba(pHba);
1364 }
1365
1366 do {
1367 rmb();
1368 m = readl(pHba->post_port);
1369 if (m != EMPTY_QUEUE) {
1370 break;
1371 }
1372 if(time_after(jiffies,timeout)){
1373 printk(KERN_WARNING"Timeout waiting for message!\n");
1374 return -ETIMEDOUT;
1375 }
a9a3047d 1376 schedule_timeout_uninterruptible(1);
1da177e4
LT
1377 } while (m == EMPTY_QUEUE);
1378
67af2b06 1379 status = dma_alloc_coherent(&pHba->pDev->dev, 4, &addr, GFP_KERNEL);
1da177e4
LT
1380 if(status == NULL) {
1381 adpt_send_nop(pHba, m);
1382 printk(KERN_ERR"IOP reset failed - no free memory.\n");
1383 return -ENOMEM;
1384 }
67af2b06 1385 memset(status,0,4);
1da177e4
LT
1386
1387 msg[0]=EIGHT_WORD_MSG_SIZE|SGL_OFFSET_0;
1388 msg[1]=I2O_CMD_ADAPTER_RESET<<24|HOST_TID<<12|ADAPTER_TID;
1389 msg[2]=0;
1390 msg[3]=0;
1391 msg[4]=0;
1392 msg[5]=0;
67af2b06
MS
1393 msg[6]=dma_low(addr);
1394 msg[7]=dma_high(addr);
1da177e4
LT
1395
1396 memcpy_toio(pHba->msg_addr_virt+m, msg, sizeof(msg));
1397 wmb();
1398 writel(m, pHba->post_port);
1399 wmb();
1400
1401 while(*status == 0){
1402 if(time_after(jiffies,timeout)){
1403 printk(KERN_WARNING"%s: IOP Reset Timeout\n",pHba->name);
67af2b06
MS
1404 /* We lose 4 bytes of "status" here, but we cannot
1405 free these because controller may awake and corrupt
1406 those bytes at any time */
1407 /* dma_free_coherent(&pHba->pDev->dev, 4, buf, addr); */
1da177e4
LT
1408 return -ETIMEDOUT;
1409 }
1410 rmb();
a9a3047d 1411 schedule_timeout_uninterruptible(1);
1da177e4
LT
1412 }
1413
1414 if(*status == 0x01 /*I2O_EXEC_IOP_RESET_IN_PROGRESS*/) {
1415 PDEBUG("%s: Reset in progress...\n", pHba->name);
1416 // Here we wait for message frame to become available
1417 // indicated that reset has finished
1418 do {
1419 rmb();
1420 m = readl(pHba->post_port);
1421 if (m != EMPTY_QUEUE) {
1422 break;
1423 }
1424 if(time_after(jiffies,timeout)){
1425 printk(KERN_ERR "%s:Timeout waiting for IOP Reset.\n",pHba->name);
67af2b06
MS
1426 /* We lose 4 bytes of "status" here, but we
1427 cannot free these because controller may
1428 awake and corrupt those bytes at any time */
1429 /* dma_free_coherent(&pHba->pDev->dev, 4, buf, addr); */
1da177e4
LT
1430 return -ETIMEDOUT;
1431 }
a9a3047d 1432 schedule_timeout_uninterruptible(1);
1da177e4
LT
1433 } while (m == EMPTY_QUEUE);
1434 // Flush the offset
1435 adpt_send_nop(pHba, m);
1436 }
1437 adpt_i2o_status_get(pHba);
1438 if(*status == 0x02 ||
1439 pHba->status_block->iop_state != ADAPTER_STATE_RESET) {
1440 printk(KERN_WARNING"%s: Reset reject, trying to clear\n",
1441 pHba->name);
1442 } else {
1443 PDEBUG("%s: Reset completed.\n", pHba->name);
1444 }
1445
67af2b06 1446 dma_free_coherent(&pHba->pDev->dev, 4, status, addr);
1da177e4
LT
1447#ifdef UARTDELAY
1448 // This delay is to allow someone attached to the card through the debug UART to
1449 // set up the dump levels that they want before the rest of the initialization sequence
1450 adpt_delay(20000);
1451#endif
1452 return 0;
1453}
1454
1455
1456static int adpt_i2o_parse_lct(adpt_hba* pHba)
1457{
1458 int i;
1459 int max;
1460 int tid;
1461 struct i2o_device *d;
1462 i2o_lct *lct = pHba->lct;
1463 u8 bus_no = 0;
1464 s16 scsi_id;
9cb78c16 1465 u64 scsi_lun;
1da177e4
LT
1466 u32 buf[10]; // larger than 7, or 8 ...
1467 struct adpt_device* pDev;
1468
1469 if (lct == NULL) {
1470 printk(KERN_ERR "%s: LCT is empty???\n",pHba->name);
1471 return -1;
1472 }
1473
1474 max = lct->table_size;
1475 max -= 3;
1476 max /= 9;
1477
1478 for(i=0;i<max;i++) {
1479 if( lct->lct_entry[i].user_tid != 0xfff){
1480 /*
1481 * If we have hidden devices, we need to inform the upper layers about
1482 * the possible maximum id reference to handle device access when
1483 * an array is disassembled. This code has no other purpose but to
1484 * allow us future access to devices that are currently hidden
1485 * behind arrays, hotspares or have not been configured (JBOD mode).
1486 */
1487 if( lct->lct_entry[i].class_id != I2O_CLASS_RANDOM_BLOCK_STORAGE &&
1488 lct->lct_entry[i].class_id != I2O_CLASS_SCSI_PERIPHERAL &&
1489 lct->lct_entry[i].class_id != I2O_CLASS_FIBRE_CHANNEL_PERIPHERAL ){
1490 continue;
1491 }
1492 tid = lct->lct_entry[i].tid;
1493 // I2O_DPT_DEVICE_INFO_GROUP_NO;
1494 if(adpt_i2o_query_scalar(pHba, tid, 0x8000, -1, buf, 32)<0) {
1495 continue;
1496 }
1497 bus_no = buf[0]>>16;
1498 scsi_id = buf[1];
9cb78c16 1499 scsi_lun = scsilun_to_int((struct scsi_lun *)&buf[2]);
1da177e4
LT
1500 if(bus_no >= MAX_CHANNEL) { // Something wrong skip it
1501 printk(KERN_WARNING"%s: Channel number %d out of range \n", pHba->name, bus_no);
1502 continue;
1503 }
1504 if (scsi_id >= MAX_ID){
1505 printk(KERN_WARNING"%s: SCSI ID %d out of range \n", pHba->name, bus_no);
1506 continue;
1507 }
1508 if(bus_no > pHba->top_scsi_channel){
1509 pHba->top_scsi_channel = bus_no;
1510 }
1511 if(scsi_id > pHba->top_scsi_id){
1512 pHba->top_scsi_id = scsi_id;
1513 }
1514 if(scsi_lun > pHba->top_scsi_lun){
1515 pHba->top_scsi_lun = scsi_lun;
1516 }
1517 continue;
1518 }
5cbded58 1519 d = kmalloc(sizeof(struct i2o_device), GFP_KERNEL);
1da177e4
LT
1520 if(d==NULL)
1521 {
1522 printk(KERN_CRIT"%s: Out of memory for I2O device data.\n",pHba->name);
1523 return -ENOMEM;
1524 }
1525
1c2fb3f3 1526 d->controller = pHba;
1da177e4
LT
1527 d->next = NULL;
1528
1529 memcpy(&d->lct_data, &lct->lct_entry[i], sizeof(i2o_lct_entry));
1530
1531 d->flags = 0;
1532 tid = d->lct_data.tid;
1533 adpt_i2o_report_hba_unit(pHba, d);
1534 adpt_i2o_install_device(pHba, d);
1535 }
1536 bus_no = 0;
1537 for(d = pHba->devices; d ; d = d->next) {
1538 if(d->lct_data.class_id == I2O_CLASS_BUS_ADAPTER_PORT ||
1539 d->lct_data.class_id == I2O_CLASS_FIBRE_CHANNEL_PORT){
1540 tid = d->lct_data.tid;
1541 // TODO get the bus_no from hrt-but for now they are in order
1542 //bus_no =
1543 if(bus_no > pHba->top_scsi_channel){
1544 pHba->top_scsi_channel = bus_no;
1545 }
1546 pHba->channel[bus_no].type = d->lct_data.class_id;
1547 pHba->channel[bus_no].tid = tid;
1548 if(adpt_i2o_query_scalar(pHba, tid, 0x0200, -1, buf, 28)>=0)
1549 {
1550 pHba->channel[bus_no].scsi_id = buf[1];
1551 PDEBUG("Bus %d - SCSI ID %d.\n", bus_no, buf[1]);
1552 }
1553 // TODO remove - this is just until we get from hrt
1554 bus_no++;
1555 if(bus_no >= MAX_CHANNEL) { // Something wrong skip it
1556 printk(KERN_WARNING"%s: Channel number %d out of range - LCT\n", pHba->name, bus_no);
1557 break;
1558 }
1559 }
1560 }
1561
1562 // Setup adpt_device table
1563 for(d = pHba->devices; d ; d = d->next) {
1564 if(d->lct_data.class_id == I2O_CLASS_RANDOM_BLOCK_STORAGE ||
1565 d->lct_data.class_id == I2O_CLASS_SCSI_PERIPHERAL ||
1566 d->lct_data.class_id == I2O_CLASS_FIBRE_CHANNEL_PERIPHERAL ){
1567
1568 tid = d->lct_data.tid;
1569 scsi_id = -1;
1570 // I2O_DPT_DEVICE_INFO_GROUP_NO;
1571 if(adpt_i2o_query_scalar(pHba, tid, 0x8000, -1, buf, 32)>=0) {
1572 bus_no = buf[0]>>16;
1573 scsi_id = buf[1];
9cb78c16 1574 scsi_lun = scsilun_to_int((struct scsi_lun *)&buf[2]);
1da177e4
LT
1575 if(bus_no >= MAX_CHANNEL) { // Something wrong skip it
1576 continue;
1577 }
1578 if (scsi_id >= MAX_ID) {
1579 continue;
1580 }
1581 if( pHba->channel[bus_no].device[scsi_id] == NULL){
ab552204 1582 pDev = kzalloc(sizeof(struct adpt_device),GFP_KERNEL);
1da177e4
LT
1583 if(pDev == NULL) {
1584 return -ENOMEM;
1585 }
1586 pHba->channel[bus_no].device[scsi_id] = pDev;
1da177e4
LT
1587 } else {
1588 for( pDev = pHba->channel[bus_no].device[scsi_id];
1589 pDev->next_lun; pDev = pDev->next_lun){
1590 }
ab552204 1591 pDev->next_lun = kzalloc(sizeof(struct adpt_device),GFP_KERNEL);
1da177e4
LT
1592 if(pDev->next_lun == NULL) {
1593 return -ENOMEM;
1594 }
1da177e4
LT
1595 pDev = pDev->next_lun;
1596 }
1597 pDev->tid = tid;
1598 pDev->scsi_channel = bus_no;
1599 pDev->scsi_id = scsi_id;
1600 pDev->scsi_lun = scsi_lun;
1601 pDev->pI2o_dev = d;
1602 d->owner = pDev;
1603 pDev->type = (buf[0])&0xff;
1604 pDev->flags = (buf[0]>>8)&0xff;
1605 if(scsi_id > pHba->top_scsi_id){
1606 pHba->top_scsi_id = scsi_id;
1607 }
1608 if(scsi_lun > pHba->top_scsi_lun){
1609 pHba->top_scsi_lun = scsi_lun;
1610 }
1611 }
1612 if(scsi_id == -1){
1613 printk(KERN_WARNING"Could not find SCSI ID for %s\n",
1614 d->lct_data.identity_tag);
1615 }
1616 }
1617 }
1618 return 0;
1619}
1620
1621
1622/*
1623 * Each I2O controller has a chain of devices on it - these match
1624 * the useful parts of the LCT of the board.
1625 */
1626
1627static int adpt_i2o_install_device(adpt_hba* pHba, struct i2o_device *d)
1628{
0b950672 1629 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
1630 d->controller=pHba;
1631 d->owner=NULL;
1632 d->next=pHba->devices;
1633 d->prev=NULL;
1634 if (pHba->devices != NULL){
1635 pHba->devices->prev=d;
1636 }
1637 pHba->devices=d;
1638 *d->dev_name = 0;
1639
0b950672 1640 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
1641 return 0;
1642}
1643
1644static int adpt_open(struct inode *inode, struct file *file)
1645{
1646 int minor;
1647 adpt_hba* pHba;
1648
c45d15d2 1649 mutex_lock(&adpt_mutex);
1da177e4
LT
1650 //TODO check for root access
1651 //
1652 minor = iminor(inode);
1653 if (minor >= hba_count) {
c45d15d2 1654 mutex_unlock(&adpt_mutex);
1da177e4
LT
1655 return -ENXIO;
1656 }
0b950672 1657 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
1658 for (pHba = hba_chain; pHba; pHba = pHba->next) {
1659 if (pHba->unit == minor) {
1660 break; /* found adapter */
1661 }
1662 }
1663 if (pHba == NULL) {
0b950672 1664 mutex_unlock(&adpt_configuration_lock);
c45d15d2 1665 mutex_unlock(&adpt_mutex);
1da177e4
LT
1666 return -ENXIO;
1667 }
1668
1669// if(pHba->in_use){
0b950672 1670 // mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
1671// return -EBUSY;
1672// }
1673
1674 pHba->in_use = 1;
0b950672 1675 mutex_unlock(&adpt_configuration_lock);
c45d15d2 1676 mutex_unlock(&adpt_mutex);
1da177e4
LT
1677
1678 return 0;
1679}
1680
1681static int adpt_close(struct inode *inode, struct file *file)
1682{
1683 int minor;
1684 adpt_hba* pHba;
1685
1686 minor = iminor(inode);
1687 if (minor >= hba_count) {
1688 return -ENXIO;
1689 }
0b950672 1690 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
1691 for (pHba = hba_chain; pHba; pHba = pHba->next) {
1692 if (pHba->unit == minor) {
1693 break; /* found adapter */
1694 }
1695 }
0b950672 1696 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
1697 if (pHba == NULL) {
1698 return -ENXIO;
1699 }
1700
1701 pHba->in_use = 0;
1702
1703 return 0;
1704}
1705
1706
1707static int adpt_i2o_passthru(adpt_hba* pHba, u32 __user *arg)
1708{
1709 u32 msg[MAX_MESSAGE_SIZE];
1710 u32* reply = NULL;
1711 u32 size = 0;
1712 u32 reply_size = 0;
1713 u32 __user *user_msg = arg;
1714 u32 __user * user_reply = NULL;
1715 void *sg_list[pHba->sg_tablesize];
1716 u32 sg_offset = 0;
1717 u32 sg_count = 0;
1718 int sg_index = 0;
1719 u32 i = 0;
1720 u32 rcode = 0;
1721 void *p = NULL;
67af2b06 1722 dma_addr_t addr;
1da177e4
LT
1723 ulong flags = 0;
1724
1725 memset(&msg, 0, MAX_MESSAGE_SIZE*4);
1726 // get user msg size in u32s
1727 if(get_user(size, &user_msg[0])){
1728 return -EFAULT;
1729 }
1730 size = size>>16;
1731
1732 user_reply = &user_msg[size];
1733 if(size > MAX_MESSAGE_SIZE){
1734 return -EFAULT;
1735 }
1736 size *= 4; // Convert to bytes
1737
1738 /* Copy in the user's I2O command */
1739 if(copy_from_user(msg, user_msg, size)) {
1740 return -EFAULT;
1741 }
1742 get_user(reply_size, &user_reply[0]);
1743 reply_size = reply_size>>16;
1744 if(reply_size > REPLY_FRAME_SIZE){
1745 reply_size = REPLY_FRAME_SIZE;
1746 }
1747 reply_size *= 4;
ab552204 1748 reply = kzalloc(REPLY_FRAME_SIZE*4, GFP_KERNEL);
1da177e4
LT
1749 if(reply == NULL) {
1750 printk(KERN_WARNING"%s: Could not allocate reply buffer\n",pHba->name);
1751 return -ENOMEM;
1752 }
1da177e4
LT
1753 sg_offset = (msg[0]>>4)&0xf;
1754 msg[2] = 0x40000000; // IOCTL context
62ac5aed 1755 msg[3] = adpt_ioctl_to_context(pHba, reply);
021e2927
QL
1756 if (msg[3] == (u32)-1) {
1757 kfree(reply);
62ac5aed 1758 return -EBUSY;
021e2927 1759 }
62ac5aed 1760
1da177e4
LT
1761 memset(sg_list,0, sizeof(sg_list[0])*pHba->sg_tablesize);
1762 if(sg_offset) {
62ac5aed 1763 // TODO add 64 bit API
1da177e4
LT
1764 struct sg_simple_element *sg = (struct sg_simple_element*) (msg+sg_offset);
1765 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element);
1766 if (sg_count > pHba->sg_tablesize){
1767 printk(KERN_DEBUG"%s:IOCTL SG List too large (%u)\n", pHba->name,sg_count);
1768 kfree (reply);
1769 return -EINVAL;
1770 }
1771
1772 for(i = 0; i < sg_count; i++) {
1773 int sg_size;
1774
1775 if (!(sg[i].flag_count & 0x10000000 /*I2O_SGL_FLAGS_SIMPLE_ADDRESS_ELEMENT*/)) {
1776 printk(KERN_DEBUG"%s:Bad SG element %d - not simple (%x)\n",pHba->name,i, sg[i].flag_count);
1777 rcode = -EINVAL;
1778 goto cleanup;
1779 }
1780 sg_size = sg[i].flag_count & 0xffffff;
1781 /* Allocate memory for the transfer */
67af2b06 1782 p = dma_alloc_coherent(&pHba->pDev->dev, sg_size, &addr, GFP_KERNEL);
1da177e4
LT
1783 if(!p) {
1784 printk(KERN_DEBUG"%s: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
1785 pHba->name,sg_size,i,sg_count);
1786 rcode = -ENOMEM;
1787 goto cleanup;
1788 }
1789 sg_list[sg_index++] = p; // sglist indexed with input frame, not our internal frame.
1790 /* Copy in the user's SG buffer if necessary */
1791 if(sg[i].flag_count & 0x04000000 /*I2O_SGL_FLAGS_DIR*/) {
62ac5aed
MS
1792 // sg_simple_element API is 32 bit
1793 if (copy_from_user(p,(void __user *)(ulong)sg[i].addr_bus, sg_size)) {
1da177e4
LT
1794 printk(KERN_DEBUG"%s: Could not copy SG buf %d FROM user\n",pHba->name,i);
1795 rcode = -EFAULT;
1796 goto cleanup;
1797 }
1798 }
62ac5aed
MS
1799 /* sg_simple_element API is 32 bit, but addr < 4GB */
1800 sg[i].addr_bus = addr;
1da177e4
LT
1801 }
1802 }
1803
1804 do {
1f8c88c3
HR
1805 /*
1806 * Stop any new commands from enterring the
1807 * controller while processing the ioctl
1808 */
1809 if (pHba->host) {
1810 scsi_block_requests(pHba->host);
1da177e4 1811 spin_lock_irqsave(pHba->host->host_lock, flags);
1f8c88c3 1812 }
1da177e4
LT
1813 rcode = adpt_i2o_post_wait(pHba, msg, size, FOREVER);
1814 if (rcode != 0)
1815 printk("adpt_i2o_passthru: post wait failed %d %p\n",
1816 rcode, reply);
1f8c88c3 1817 if (pHba->host) {
1da177e4 1818 spin_unlock_irqrestore(pHba->host->host_lock, flags);
1f8c88c3
HR
1819 scsi_unblock_requests(pHba->host);
1820 }
1821 } while (rcode == -ETIMEDOUT);
1da177e4
LT
1822
1823 if(rcode){
1824 goto cleanup;
1825 }
1826
1827 if(sg_offset) {
1828 /* Copy back the Scatter Gather buffers back to user space */
1829 u32 j;
62ac5aed 1830 // TODO add 64 bit API
1da177e4
LT
1831 struct sg_simple_element* sg;
1832 int sg_size;
1833
1834 // re-acquire the original message to handle correctly the sg copy operation
1835 memset(&msg, 0, MAX_MESSAGE_SIZE*4);
1836 // get user msg size in u32s
1837 if(get_user(size, &user_msg[0])){
1838 rcode = -EFAULT;
1839 goto cleanup;
1840 }
1841 size = size>>16;
1842 size *= 4;
ef7562b7 1843 if (size > MAX_MESSAGE_SIZE) {
aefba418 1844 rcode = -EINVAL;
ef7562b7
AC
1845 goto cleanup;
1846 }
1da177e4
LT
1847 /* Copy in the user's I2O command */
1848 if (copy_from_user (msg, user_msg, size)) {
1849 rcode = -EFAULT;
1850 goto cleanup;
1851 }
1852 sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element);
1853
62ac5aed 1854 // TODO add 64 bit API
1da177e4
LT
1855 sg = (struct sg_simple_element*)(msg + sg_offset);
1856 for (j = 0; j < sg_count; j++) {
1857 /* Copy out the SG list to user's buffer if necessary */
1858 if(! (sg[j].flag_count & 0x4000000 /*I2O_SGL_FLAGS_DIR*/)) {
1859 sg_size = sg[j].flag_count & 0xffffff;
62ac5aed
MS
1860 // sg_simple_element API is 32 bit
1861 if (copy_to_user((void __user *)(ulong)sg[j].addr_bus,sg_list[j], sg_size)) {
1da177e4
LT
1862 printk(KERN_WARNING"%s: Could not copy %p TO user %x\n",pHba->name, sg_list[j], sg[j].addr_bus);
1863 rcode = -EFAULT;
1864 goto cleanup;
1865 }
1866 }
1867 }
1868 }
1869
1870 /* Copy back the reply to user space */
1871 if (reply_size) {
1872 // we wrote our own values for context - now restore the user supplied ones
1873 if(copy_from_user(reply+2, user_msg+2, sizeof(u32)*2)) {
1874 printk(KERN_WARNING"%s: Could not copy message context FROM user\n",pHba->name);
1875 rcode = -EFAULT;
1876 }
1877 if(copy_to_user(user_reply, reply, reply_size)) {
1878 printk(KERN_WARNING"%s: Could not copy reply TO user\n",pHba->name);
1879 rcode = -EFAULT;
1880 }
1881 }
1882
1883
1884cleanup:
67af2b06
MS
1885 if (rcode != -ETIME && rcode != -EINTR) {
1886 struct sg_simple_element *sg =
1887 (struct sg_simple_element*) (msg +sg_offset);
1da177e4 1888 kfree (reply);
67af2b06
MS
1889 while(sg_index) {
1890 if(sg_list[--sg_index]) {
1891 dma_free_coherent(&pHba->pDev->dev,
1892 sg[sg_index].flag_count & 0xffffff,
1893 sg_list[sg_index],
1894 sg[sg_index].addr_bus);
1895 }
1da177e4
LT
1896 }
1897 }
1898 return rcode;
1899}
1900
1da177e4
LT
1901#if defined __ia64__
1902static void adpt_ia64_info(sysInfo_S* si)
1903{
1904 // This is all the info we need for now
1905 // We will add more info as our new
1906 // managmenent utility requires it
1907 si->processorType = PROC_IA64;
1908}
1909#endif
1910
1da177e4
LT
1911#if defined __sparc__
1912static void adpt_sparc_info(sysInfo_S* si)
1913{
1914 // This is all the info we need for now
1915 // We will add more info as our new
1916 // managmenent utility requires it
1917 si->processorType = PROC_ULTRASPARC;
1918}
1919#endif
1da177e4
LT
1920#if defined __alpha__
1921static void adpt_alpha_info(sysInfo_S* si)
1922{
1923 // This is all the info we need for now
1924 // We will add more info as our new
1925 // managmenent utility requires it
1926 si->processorType = PROC_ALPHA;
1927}
1928#endif
1929
1930#if defined __i386__
ba3e127e
BG
1931
1932#include <uapi/asm/vm86.h>
1933
1da177e4
LT
1934static void adpt_i386_info(sysInfo_S* si)
1935{
1936 // This is all the info we need for now
1937 // We will add more info as our new
1938 // managmenent utility requires it
1939 switch (boot_cpu_data.x86) {
1940 case CPU_386:
1941 si->processorType = PROC_386;
1942 break;
1943 case CPU_486:
1944 si->processorType = PROC_486;
1945 break;
1946 case CPU_586:
1947 si->processorType = PROC_PENTIUM;
1948 break;
1949 default: // Just in case
1950 si->processorType = PROC_PENTIUM;
1951 break;
1952 }
1953}
8b2cc917
AM
1954#endif
1955
1956/*
1957 * This routine returns information about the system. This does not effect
1958 * any logic and if the info is wrong - it doesn't matter.
1959 */
1da177e4 1960
8b2cc917
AM
1961/* Get all the info we can not get from kernel services */
1962static int adpt_system_info(void __user *buffer)
1963{
1964 sysInfo_S si;
1965
1966 memset(&si, 0, sizeof(si));
1967
1968 si.osType = OS_LINUX;
1969 si.osMajorVersion = 0;
1970 si.osMinorVersion = 0;
1971 si.osRevision = 0;
1972 si.busType = SI_PCI_BUS;
1973 si.processorFamily = DPTI_sig.dsProcessorFamily;
1974
1975#if defined __i386__
1976 adpt_i386_info(&si);
1977#elif defined (__ia64__)
1978 adpt_ia64_info(&si);
1979#elif defined(__sparc__)
1980 adpt_sparc_info(&si);
1981#elif defined (__alpha__)
1982 adpt_alpha_info(&si);
1983#else
1984 si.processorType = 0xff ;
1da177e4 1985#endif
8b2cc917
AM
1986 if (copy_to_user(buffer, &si, sizeof(si))){
1987 printk(KERN_WARNING"dpti: Could not copy buffer TO user\n");
1988 return -EFAULT;
1989 }
1da177e4 1990
8b2cc917
AM
1991 return 0;
1992}
1da177e4 1993
f4927c45 1994static int adpt_ioctl(struct inode *inode, struct file *file, uint cmd, ulong arg)
1da177e4
LT
1995{
1996 int minor;
1997 int error = 0;
1998 adpt_hba* pHba;
1999 ulong flags = 0;
2000 void __user *argp = (void __user *)arg;
2001
2002 minor = iminor(inode);
2003 if (minor >= DPTI_MAX_HBA){
2004 return -ENXIO;
2005 }
0b950672 2006 mutex_lock(&adpt_configuration_lock);
1da177e4
LT
2007 for (pHba = hba_chain; pHba; pHba = pHba->next) {
2008 if (pHba->unit == minor) {
2009 break; /* found adapter */
2010 }
2011 }
0b950672 2012 mutex_unlock(&adpt_configuration_lock);
1da177e4
LT
2013 if(pHba == NULL){
2014 return -ENXIO;
2015 }
2016
a9a3047d
NA
2017 while((volatile u32) pHba->state & DPTI_STATE_RESET )
2018 schedule_timeout_uninterruptible(2);
1da177e4
LT
2019
2020 switch (cmd) {
2021 // TODO: handle 3 cases
2022 case DPT_SIGNATURE:
2023 if (copy_to_user(argp, &DPTI_sig, sizeof(DPTI_sig))) {
2024 return -EFAULT;
2025 }
2026 break;
2027 case I2OUSRCMD:
2028 return adpt_i2o_passthru(pHba, argp);
2029
2030 case DPT_CTRLINFO:{
2031 drvrHBAinfo_S HbaInfo;
2032
2033#define FLG_OSD_PCI_VALID 0x0001
2034#define FLG_OSD_DMA 0x0002
2035#define FLG_OSD_I2O 0x0004
2036 memset(&HbaInfo, 0, sizeof(HbaInfo));
2037 HbaInfo.drvrHBAnum = pHba->unit;
2038 HbaInfo.baseAddr = (ulong) pHba->base_addr_phys;
2039 HbaInfo.blinkState = adpt_read_blink_led(pHba);
2040 HbaInfo.pciBusNum = pHba->pDev->bus->number;
2041 HbaInfo.pciDeviceNum=PCI_SLOT(pHba->pDev->devfn);
2042 HbaInfo.Interrupt = pHba->pDev->irq;
2043 HbaInfo.hbaFlags = FLG_OSD_PCI_VALID | FLG_OSD_DMA | FLG_OSD_I2O;
2044 if(copy_to_user(argp, &HbaInfo, sizeof(HbaInfo))){
2045 printk(KERN_WARNING"%s: Could not copy HbaInfo TO user\n",pHba->name);
2046 return -EFAULT;
2047 }
2048 break;
2049 }
2050 case DPT_SYSINFO:
2051 return adpt_system_info(argp);
2052 case DPT_BLINKLED:{
2053 u32 value;
2054 value = (u32)adpt_read_blink_led(pHba);
2055 if (copy_to_user(argp, &value, sizeof(value))) {
2056 return -EFAULT;
2057 }
2058 break;
2059 }
2060 case I2ORESETCMD:
2061 if(pHba->host)
2062 spin_lock_irqsave(pHba->host->host_lock, flags);
2063 adpt_hba_reset(pHba);
2064 if(pHba->host)
2065 spin_unlock_irqrestore(pHba->host->host_lock, flags);
2066 break;
2067 case I2ORESCANCMD:
2068 adpt_rescan(pHba);
2069 break;
2070 default:
2071 return -EINVAL;
2072 }
2073
2074 return error;
2075}
2076
f4927c45
AB
2077static long adpt_unlocked_ioctl(struct file *file, uint cmd, ulong arg)
2078{
2079 struct inode *inode;
2080 long ret;
2081
496ad9aa 2082 inode = file_inode(file);
f4927c45 2083
c45d15d2 2084 mutex_lock(&adpt_mutex);
f4927c45 2085 ret = adpt_ioctl(inode, file, cmd, arg);
c45d15d2 2086 mutex_unlock(&adpt_mutex);
f4927c45
AB
2087
2088 return ret;
2089}
2090
62ac5aed
MS
2091#ifdef CONFIG_COMPAT
2092static long compat_adpt_ioctl(struct file *file,
2093 unsigned int cmd, unsigned long arg)
2094{
2095 struct inode *inode;
2096 long ret;
2097
496ad9aa 2098 inode = file_inode(file);
62ac5aed 2099
c45d15d2 2100 mutex_lock(&adpt_mutex);
62ac5aed
MS
2101
2102 switch(cmd) {
2103 case DPT_SIGNATURE:
2104 case I2OUSRCMD:
2105 case DPT_CTRLINFO:
2106 case DPT_SYSINFO:
2107 case DPT_BLINKLED:
2108 case I2ORESETCMD:
2109 case I2ORESCANCMD:
2110 case (DPT_TARGET_BUSY & 0xFFFF):
2111 case DPT_TARGET_BUSY:
2112 ret = adpt_ioctl(inode, file, cmd, arg);
2113 break;
2114 default:
2115 ret = -ENOIOCTLCMD;
2116 }
2117
c45d15d2 2118 mutex_unlock(&adpt_mutex);
62ac5aed
MS
2119
2120 return ret;
2121}
2122#endif
1da177e4 2123
7d12e780 2124static irqreturn_t adpt_isr(int irq, void *dev_id)
1da177e4
LT
2125{
2126 struct scsi_cmnd* cmd;
2127 adpt_hba* pHba = dev_id;
2128 u32 m;
1c2fb3f3 2129 void __iomem *reply;
1da177e4
LT
2130 u32 status=0;
2131 u32 context;
2132 ulong flags = 0;
2133 int handled = 0;
2134
2135 if (pHba == NULL){
2136 printk(KERN_WARNING"adpt_isr: NULL dev_id\n");
2137 return IRQ_NONE;
2138 }
2139 if(pHba->host)
2140 spin_lock_irqsave(pHba->host->host_lock, flags);
2141
2142 while( readl(pHba->irq_mask) & I2O_INTERRUPT_PENDING_B) {
2143 m = readl(pHba->reply_port);
2144 if(m == EMPTY_QUEUE){
2145 // Try twice then give up
2146 rmb();
2147 m = readl(pHba->reply_port);
2148 if(m == EMPTY_QUEUE){
2149 // This really should not happen
2150 printk(KERN_ERR"dpti: Could not get reply frame\n");
2151 goto out;
2152 }
2153 }
67af2b06
MS
2154 if (pHba->reply_pool_pa <= m &&
2155 m < pHba->reply_pool_pa +
2156 (pHba->reply_fifo_size * REPLY_FRAME_SIZE * 4)) {
2157 reply = (u8 *)pHba->reply_pool +
2158 (m - pHba->reply_pool_pa);
2159 } else {
2160 /* Ick, we should *never* be here */
2161 printk(KERN_ERR "dpti: reply frame not from pool\n");
2162 reply = (u8 *)bus_to_virt(m);
2163 }
1da177e4
LT
2164
2165 if (readl(reply) & MSG_FAIL) {
2166 u32 old_m = readl(reply+28);
1c2fb3f3 2167 void __iomem *msg;
1da177e4
LT
2168 u32 old_context;
2169 PDEBUG("%s: Failed message\n",pHba->name);
2170 if(old_m >= 0x100000){
2171 printk(KERN_ERR"%s: Bad preserved MFA (%x)- dropping frame\n",pHba->name,old_m);
2172 writel(m,pHba->reply_port);
2173 continue;
2174 }
2175 // Transaction context is 0 in failed reply frame
1c2fb3f3 2176 msg = pHba->msg_addr_virt + old_m;
1da177e4
LT
2177 old_context = readl(msg+12);
2178 writel(old_context, reply+12);
2179 adpt_send_nop(pHba, old_m);
2180 }
2181 context = readl(reply+8);
2182 if(context & 0x40000000){ // IOCTL
62ac5aed 2183 void *p = adpt_ioctl_from_context(pHba, readl(reply+12));
1c2fb3f3
BB
2184 if( p != NULL) {
2185 memcpy_fromio(p, reply, REPLY_FRAME_SIZE * 4);
1da177e4
LT
2186 }
2187 // All IOCTLs will also be post wait
2188 }
2189 if(context & 0x80000000){ // Post wait message
2190 status = readl(reply+16);
2191 if(status >> 24){
2192 status &= 0xffff; /* Get detail status */
2193 } else {
2194 status = I2O_POST_WAIT_OK;
2195 }
2196 if(!(context & 0x40000000)) {
62ac5aed
MS
2197 cmd = adpt_cmd_from_context(pHba,
2198 readl(reply+12));
1da177e4
LT
2199 if(cmd != NULL) {
2200 printk(KERN_WARNING"%s: Apparent SCSI cmd in Post Wait Context - cmd=%p context=%x\n", pHba->name, cmd, context);
2201 }
2202 }
2203 adpt_i2o_post_wait_complete(context, status);
2204 } else { // SCSI message
62ac5aed 2205 cmd = adpt_cmd_from_context (pHba, readl(reply+12));
1da177e4 2206 if(cmd != NULL){
67af2b06 2207 scsi_dma_unmap(cmd);
1da177e4
LT
2208 if(cmd->serial_number != 0) { // If not timedout
2209 adpt_i2o_to_scsi(reply, cmd);
2210 }
2211 }
2212 }
2213 writel(m, pHba->reply_port);
2214 wmb();
2215 rmb();
2216 }
2217 handled = 1;
2218out: if(pHba->host)
2219 spin_unlock_irqrestore(pHba->host->host_lock, flags);
2220 return IRQ_RETVAL(handled);
2221}
2222
2223static s32 adpt_scsi_to_i2o(adpt_hba* pHba, struct scsi_cmnd* cmd, struct adpt_device* d)
2224{
2225 int i;
2226 u32 msg[MAX_MESSAGE_SIZE];
2227 u32* mptr;
62ac5aed 2228 u32* lptr;
1da177e4
LT
2229 u32 *lenptr;
2230 int direction;
2231 int scsidir;
10803de4 2232 int nseg;
1da177e4
LT
2233 u32 len;
2234 u32 reqlen;
2235 s32 rcode;
62ac5aed 2236 dma_addr_t addr;
1da177e4
LT
2237
2238 memset(msg, 0 , sizeof(msg));
10803de4 2239 len = scsi_bufflen(cmd);
1da177e4
LT
2240 direction = 0x00000000;
2241
2242 scsidir = 0x00000000; // DATA NO XFER
2243 if(len) {
2244 /*
2245 * Set SCBFlags to indicate if data is being transferred
2246 * in or out, or no data transfer
2247 * Note: Do not have to verify index is less than 0 since
2248 * cmd->cmnd[0] is an unsigned char
2249 */
2250 switch(cmd->sc_data_direction){
2251 case DMA_FROM_DEVICE:
2252 scsidir =0x40000000; // DATA IN (iop<--dev)
2253 break;
2254 case DMA_TO_DEVICE:
2255 direction=0x04000000; // SGL OUT
2256 scsidir =0x80000000; // DATA OUT (iop-->dev)
2257 break;
2258 case DMA_NONE:
2259 break;
2260 case DMA_BIDIRECTIONAL:
2261 scsidir =0x40000000; // DATA IN (iop<--dev)
2262 // Assume In - and continue;
2263 break;
2264 default:
2265 printk(KERN_WARNING"%s: scsi opcode 0x%x not supported.\n",
2266 pHba->name, cmd->cmnd[0]);
2267 cmd->result = (DID_OK <<16) | (INITIATOR_ERROR << 8);
2268 cmd->scsi_done(cmd);
2269 return 0;
2270 }
2271 }
2272 // msg[0] is set later
2273 // I2O_CMD_SCSI_EXEC
2274 msg[1] = ((0xff<<24)|(HOST_TID<<12)|d->tid);
2275 msg[2] = 0;
62ac5aed 2276 msg[3] = adpt_cmd_to_context(cmd); /* Want SCSI control block back */
1da177e4
LT
2277 // Our cards use the transaction context as the tag for queueing
2278 // Adaptec/DPT Private stuff
2279 msg[4] = I2O_CMD_SCSI_EXEC|(DPT_ORGANIZATION_ID<<16);
2280 msg[5] = d->tid;
2281 /* Direction, disconnect ok | sense data | simple queue , CDBLen */
2282 // I2O_SCB_FLAG_ENABLE_DISCONNECT |
2283 // I2O_SCB_FLAG_SIMPLE_QUEUE_TAG |
2284 // I2O_SCB_FLAG_SENSE_DATA_IN_MESSAGE;
2285 msg[6] = scsidir|0x20a00000|cmd->cmd_len;
2286
2287 mptr=msg+7;
2288
2289 // Write SCSI command into the message - always 16 byte block
2290 memset(mptr, 0, 16);
2291 memcpy(mptr, cmd->cmnd, cmd->cmd_len);
2292 mptr+=4;
2293 lenptr=mptr++; /* Remember me - fill in when we know */
62ac5aed
MS
2294 if (dpt_dma64(pHba)) {
2295 reqlen = 16; // SINGLE SGE
2296 *mptr++ = (0x7C<<24)+(2<<16)+0x02; /* Enable 64 bit */
2297 *mptr++ = 1 << PAGE_SHIFT;
2298 } else {
2299 reqlen = 14; // SINGLE SGE
2300 }
1da177e4 2301 /* Now fill in the SGList and command */
1da177e4 2302
10803de4
FT
2303 nseg = scsi_dma_map(cmd);
2304 BUG_ON(nseg < 0);
2305 if (nseg) {
2306 struct scatterlist *sg;
1da177e4
LT
2307
2308 len = 0;
10803de4 2309 scsi_for_each_sg(cmd, sg, nseg, i) {
62ac5aed 2310 lptr = mptr;
1da177e4
LT
2311 *mptr++ = direction|0x10000000|sg_dma_len(sg);
2312 len+=sg_dma_len(sg);
62ac5aed
MS
2313 addr = sg_dma_address(sg);
2314 *mptr++ = dma_low(addr);
2315 if (dpt_dma64(pHba))
2316 *mptr++ = dma_high(addr);
10803de4
FT
2317 /* Make this an end of list */
2318 if (i == nseg - 1)
62ac5aed 2319 *lptr = direction|0xD0000000|sg_dma_len(sg);
1da177e4 2320 }
1da177e4
LT
2321 reqlen = mptr - msg;
2322 *lenptr = len;
2323
2324 if(cmd->underflow && len != cmd->underflow){
2325 printk(KERN_WARNING"Cmd len %08X Cmd underflow %08X\n",
2326 len, cmd->underflow);
2327 }
2328 } else {
10803de4
FT
2329 *lenptr = len = 0;
2330 reqlen = 12;
1da177e4
LT
2331 }
2332
2333 /* Stick the headers on */
2334 msg[0] = reqlen<<16 | ((reqlen > 12) ? SGL_OFFSET_12 : SGL_OFFSET_0);
2335
2336 // Send it on it's way
2337 rcode = adpt_i2o_post_this(pHba, msg, reqlen<<2);
2338 if (rcode == 0) {
2339 return 0;
2340 }
2341 return rcode;
2342}
2343
2344
c864cb14 2345static s32 adpt_scsi_host_alloc(adpt_hba* pHba, struct scsi_host_template *sht)
24601bbc 2346{
c864cb14 2347 struct Scsi_Host *host;
24601bbc 2348
c864cb14 2349 host = scsi_host_alloc(sht, sizeof(adpt_hba*));
24601bbc 2350 if (host == NULL) {
c864cb14 2351 printk("%s: scsi_host_alloc returned NULL\n", pHba->name);
24601bbc
AM
2352 return -1;
2353 }
2354 host->hostdata[0] = (unsigned long)pHba;
2355 pHba->host = host;
2356
2357 host->irq = pHba->pDev->irq;
2358 /* no IO ports, so don't have to set host->io_port and
2359 * host->n_io_port
2360 */
2361 host->io_port = 0;
2362 host->n_io_port = 0;
2363 /* see comments in scsi_host.h */
2364 host->max_id = 16;
2365 host->max_lun = 256;
2366 host->max_channel = pHba->top_scsi_channel + 1;
2367 host->cmd_per_lun = 1;
67af2b06 2368 host->unique_id = (u32)sys_tbl_pa + pHba->unit;
24601bbc
AM
2369 host->sg_tablesize = pHba->sg_tablesize;
2370 host->can_queue = pHba->post_fifo_size;
64bdcbc4 2371 host->use_cmd_list = 1;
24601bbc
AM
2372
2373 return 0;
2374}
2375
2376
1c2fb3f3 2377static s32 adpt_i2o_to_scsi(void __iomem *reply, struct scsi_cmnd* cmd)
1da177e4
LT
2378{
2379 adpt_hba* pHba;
2380 u32 hba_status;
2381 u32 dev_status;
2382 u32 reply_flags = readl(reply) & 0xff00; // Leave it shifted up 8 bits
2383 // I know this would look cleaner if I just read bytes
2384 // but the model I have been using for all the rest of the
2385 // io is in 4 byte words - so I keep that model
2386 u16 detailed_status = readl(reply+16) &0xffff;
2387 dev_status = (detailed_status & 0xff);
2388 hba_status = detailed_status >> 8;
2389
2390 // calculate resid for sg
df81d237 2391 scsi_set_resid(cmd, scsi_bufflen(cmd) - readl(reply+20));
1da177e4
LT
2392
2393 pHba = (adpt_hba*) cmd->device->host->hostdata[0];
2394
2395 cmd->sense_buffer[0] = '\0'; // initialize sense valid flag to false
2396
2397 if(!(reply_flags & MSG_FAIL)) {
2398 switch(detailed_status & I2O_SCSI_DSC_MASK) {
2399 case I2O_SCSI_DSC_SUCCESS:
2400 cmd->result = (DID_OK << 16);
2401 // handle underflow
df81d237 2402 if (readl(reply+20) < cmd->underflow) {
1da177e4
LT
2403 cmd->result = (DID_ERROR <<16);
2404 printk(KERN_WARNING"%s: SCSI CMD underflow\n",pHba->name);
2405 }
2406 break;
2407 case I2O_SCSI_DSC_REQUEST_ABORTED:
2408 cmd->result = (DID_ABORT << 16);
2409 break;
2410 case I2O_SCSI_DSC_PATH_INVALID:
2411 case I2O_SCSI_DSC_DEVICE_NOT_PRESENT:
2412 case I2O_SCSI_DSC_SELECTION_TIMEOUT:
2413 case I2O_SCSI_DSC_COMMAND_TIMEOUT:
2414 case I2O_SCSI_DSC_NO_ADAPTER:
2415 case I2O_SCSI_DSC_RESOURCE_UNAVAILABLE:
9cb78c16
HR
2416 printk(KERN_WARNING"%s: SCSI Timeout-Device (%d,%d,%llu) hba status=0x%x, dev status=0x%x, cmd=0x%x\n",
2417 pHba->name, (u32)cmd->device->channel, (u32)cmd->device->id, cmd->device->lun, hba_status, dev_status, cmd->cmnd[0]);
1da177e4
LT
2418 cmd->result = (DID_TIME_OUT << 16);
2419 break;
2420 case I2O_SCSI_DSC_ADAPTER_BUSY:
2421 case I2O_SCSI_DSC_BUS_BUSY:
2422 cmd->result = (DID_BUS_BUSY << 16);
2423 break;
2424 case I2O_SCSI_DSC_SCSI_BUS_RESET:
2425 case I2O_SCSI_DSC_BDR_MESSAGE_SENT:
2426 cmd->result = (DID_RESET << 16);
2427 break;
2428 case I2O_SCSI_DSC_PARITY_ERROR_FAILURE:
2429 printk(KERN_WARNING"%s: SCSI CMD parity error\n",pHba->name);
2430 cmd->result = (DID_PARITY << 16);
2431 break;
2432 case I2O_SCSI_DSC_UNABLE_TO_ABORT:
2433 case I2O_SCSI_DSC_COMPLETE_WITH_ERROR:
2434 case I2O_SCSI_DSC_UNABLE_TO_TERMINATE:
2435 case I2O_SCSI_DSC_MR_MESSAGE_RECEIVED:
2436 case I2O_SCSI_DSC_AUTOSENSE_FAILED:
2437 case I2O_SCSI_DSC_DATA_OVERRUN:
2438 case I2O_SCSI_DSC_UNEXPECTED_BUS_FREE:
2439 case I2O_SCSI_DSC_SEQUENCE_FAILURE:
2440 case I2O_SCSI_DSC_REQUEST_LENGTH_ERROR:
2441 case I2O_SCSI_DSC_PROVIDE_FAILURE:
2442 case I2O_SCSI_DSC_REQUEST_TERMINATED:
2443 case I2O_SCSI_DSC_IDE_MESSAGE_SENT:
2444 case I2O_SCSI_DSC_UNACKNOWLEDGED_EVENT:
2445 case I2O_SCSI_DSC_MESSAGE_RECEIVED:
2446 case I2O_SCSI_DSC_INVALID_CDB:
2447 case I2O_SCSI_DSC_LUN_INVALID:
2448 case I2O_SCSI_DSC_SCSI_TID_INVALID:
2449 case I2O_SCSI_DSC_FUNCTION_UNAVAILABLE:
2450 case I2O_SCSI_DSC_NO_NEXUS:
2451 case I2O_SCSI_DSC_CDB_RECEIVED:
2452 case I2O_SCSI_DSC_LUN_ALREADY_ENABLED:
2453 case I2O_SCSI_DSC_QUEUE_FROZEN:
2454 case I2O_SCSI_DSC_REQUEST_INVALID:
2455 default:
9cb78c16
HR
2456 printk(KERN_WARNING"%s: SCSI error %0x-Device(%d,%d,%llu) hba_status=0x%x, dev_status=0x%x, cmd=0x%x\n",
2457 pHba->name, detailed_status & I2O_SCSI_DSC_MASK, (u32)cmd->device->channel, (u32)cmd->device->id, cmd->device->lun,
1da177e4
LT
2458 hba_status, dev_status, cmd->cmnd[0]);
2459 cmd->result = (DID_ERROR << 16);
2460 break;
2461 }
2462
2463 // copy over the request sense data if it was a check
2464 // condition status
d814c517 2465 if (dev_status == SAM_STAT_CHECK_CONDITION) {
b80ca4f7 2466 u32 len = min(SCSI_SENSE_BUFFERSIZE, 40);
1da177e4 2467 // Copy over the sense data
1c2fb3f3 2468 memcpy_fromio(cmd->sense_buffer, (reply+28) , len);
1da177e4
LT
2469 if(cmd->sense_buffer[0] == 0x70 /* class 7 */ &&
2470 cmd->sense_buffer[2] == DATA_PROTECT ){
2471 /* This is to handle an array failed */
2472 cmd->result = (DID_TIME_OUT << 16);
9cb78c16
HR
2473 printk(KERN_WARNING"%s: SCSI Data Protect-Device (%d,%d,%llu) hba_status=0x%x, dev_status=0x%x, cmd=0x%x\n",
2474 pHba->name, (u32)cmd->device->channel, (u32)cmd->device->id, cmd->device->lun,
1da177e4
LT
2475 hba_status, dev_status, cmd->cmnd[0]);
2476
2477 }
2478 }
2479 } else {
2480 /* In this condtion we could not talk to the tid
2481 * the card rejected it. We should signal a retry
2482 * for a limitted number of retries.
2483 */
2484 cmd->result = (DID_TIME_OUT << 16);
9cb78c16
HR
2485 printk(KERN_WARNING"%s: I2O MSG_FAIL - Device (%d,%d,%llu) tid=%d, cmd=0x%x\n",
2486 pHba->name, (u32)cmd->device->channel, (u32)cmd->device->id, cmd->device->lun,
1da177e4
LT
2487 ((struct adpt_device*)(cmd->device->hostdata))->tid, cmd->cmnd[0]);
2488 }
2489
2490 cmd->result |= (dev_status);
2491
2492 if(cmd->scsi_done != NULL){
2493 cmd->scsi_done(cmd);
2494 }
2495 return cmd->result;
2496}
2497
2498
2499static s32 adpt_rescan(adpt_hba* pHba)
2500{
2501 s32 rcode;
2502 ulong flags = 0;
2503
2504 if(pHba->host)
2505 spin_lock_irqsave(pHba->host->host_lock, flags);
2506 if ((rcode=adpt_i2o_lct_get(pHba)) < 0)
2507 goto out;
2508 if ((rcode=adpt_i2o_reparse_lct(pHba)) < 0)
2509 goto out;
2510 rcode = 0;
2511out: if(pHba->host)
2512 spin_unlock_irqrestore(pHba->host->host_lock, flags);
2513 return rcode;
2514}
2515
2516
2517static s32 adpt_i2o_reparse_lct(adpt_hba* pHba)
2518{
2519 int i;
2520 int max;
2521 int tid;
2522 struct i2o_device *d;
2523 i2o_lct *lct = pHba->lct;
2524 u8 bus_no = 0;
2525 s16 scsi_id;
9cb78c16 2526 u64 scsi_lun;
1da177e4
LT
2527 u32 buf[10]; // at least 8 u32's
2528 struct adpt_device* pDev = NULL;
2529 struct i2o_device* pI2o_dev = NULL;
2530
2531 if (lct == NULL) {
2532 printk(KERN_ERR "%s: LCT is empty???\n",pHba->name);
2533 return -1;
2534 }
2535
2536 max = lct->table_size;
2537 max -= 3;
2538 max /= 9;
2539
2540 // Mark each drive as unscanned
2541 for (d = pHba->devices; d; d = d->next) {
2542 pDev =(struct adpt_device*) d->owner;
2543 if(!pDev){
2544 continue;
2545 }
2546 pDev->state |= DPTI_DEV_UNSCANNED;
2547 }
2548
2549 printk(KERN_INFO "%s: LCT has %d entries.\n", pHba->name,max);
2550
2551 for(i=0;i<max;i++) {
2552 if( lct->lct_entry[i].user_tid != 0xfff){
2553 continue;
2554 }
2555
2556 if( lct->lct_entry[i].class_id == I2O_CLASS_RANDOM_BLOCK_STORAGE ||
2557 lct->lct_entry[i].class_id == I2O_CLASS_SCSI_PERIPHERAL ||
2558 lct->lct_entry[i].class_id == I2O_CLASS_FIBRE_CHANNEL_PERIPHERAL ){
2559 tid = lct->lct_entry[i].tid;
2560 if(adpt_i2o_query_scalar(pHba, tid, 0x8000, -1, buf, 32)<0) {
2561 printk(KERN_ERR"%s: Could not query device\n",pHba->name);
2562 continue;
2563 }
2564 bus_no = buf[0]>>16;
e84d96db
DC
2565 if (bus_no >= MAX_CHANNEL) { /* Something wrong skip it */
2566 printk(KERN_WARNING
2567 "%s: Channel number %d out of range\n",
2568 pHba->name, bus_no);
2569 continue;
2570 }
2571
1da177e4 2572 scsi_id = buf[1];
9cb78c16 2573 scsi_lun = scsilun_to_int((struct scsi_lun *)&buf[2]);
1da177e4
LT
2574 pDev = pHba->channel[bus_no].device[scsi_id];
2575 /* da lun */
2576 while(pDev) {
2577 if(pDev->scsi_lun == scsi_lun) {
2578 break;
2579 }
2580 pDev = pDev->next_lun;
2581 }
2582 if(!pDev ) { // Something new add it
da2907ff
JL
2583 d = kmalloc(sizeof(struct i2o_device),
2584 GFP_ATOMIC);
1da177e4
LT
2585 if(d==NULL)
2586 {
2587 printk(KERN_CRIT "Out of memory for I2O device data.\n");
2588 return -ENOMEM;
2589 }
2590
1c2fb3f3 2591 d->controller = pHba;
1da177e4
LT
2592 d->next = NULL;
2593
2594 memcpy(&d->lct_data, &lct->lct_entry[i], sizeof(i2o_lct_entry));
2595
2596 d->flags = 0;
2597 adpt_i2o_report_hba_unit(pHba, d);
2598 adpt_i2o_install_device(pHba, d);
2599
1da177e4
LT
2600 pDev = pHba->channel[bus_no].device[scsi_id];
2601 if( pDev == NULL){
da2907ff
JL
2602 pDev =
2603 kzalloc(sizeof(struct adpt_device),
2604 GFP_ATOMIC);
1da177e4
LT
2605 if(pDev == NULL) {
2606 return -ENOMEM;
2607 }
2608 pHba->channel[bus_no].device[scsi_id] = pDev;
2609 } else {
2610 while (pDev->next_lun) {
2611 pDev = pDev->next_lun;
2612 }
da2907ff
JL
2613 pDev = pDev->next_lun =
2614 kzalloc(sizeof(struct adpt_device),
2615 GFP_ATOMIC);
1da177e4
LT
2616 if(pDev == NULL) {
2617 return -ENOMEM;
2618 }
2619 }
1da177e4
LT
2620 pDev->tid = d->lct_data.tid;
2621 pDev->scsi_channel = bus_no;
2622 pDev->scsi_id = scsi_id;
2623 pDev->scsi_lun = scsi_lun;
2624 pDev->pI2o_dev = d;
2625 d->owner = pDev;
2626 pDev->type = (buf[0])&0xff;
2627 pDev->flags = (buf[0]>>8)&0xff;
2628 // Too late, SCSI system has made up it's mind, but what the hey ...
2629 if(scsi_id > pHba->top_scsi_id){
2630 pHba->top_scsi_id = scsi_id;
2631 }
2632 if(scsi_lun > pHba->top_scsi_lun){
2633 pHba->top_scsi_lun = scsi_lun;
2634 }
2635 continue;
2636 } // end of new i2o device
2637
2638 // We found an old device - check it
2639 while(pDev) {
2640 if(pDev->scsi_lun == scsi_lun) {
2641 if(!scsi_device_online(pDev->pScsi_dev)) {
9cb78c16 2642 printk(KERN_WARNING"%s: Setting device (%d,%d,%llu) back online\n",
1da177e4
LT
2643 pHba->name,bus_no,scsi_id,scsi_lun);
2644 if (pDev->pScsi_dev) {
2645 scsi_device_set_state(pDev->pScsi_dev, SDEV_RUNNING);
2646 }
2647 }
2648 d = pDev->pI2o_dev;
2649 if(d->lct_data.tid != tid) { // something changed
2650 pDev->tid = tid;
2651 memcpy(&d->lct_data, &lct->lct_entry[i], sizeof(i2o_lct_entry));
2652 if (pDev->pScsi_dev) {
2653 pDev->pScsi_dev->changed = TRUE;
2654 pDev->pScsi_dev->removable = TRUE;
2655 }
2656 }
2657 // Found it - mark it scanned
2658 pDev->state = DPTI_DEV_ONLINE;
2659 break;
2660 }
2661 pDev = pDev->next_lun;
2662 }
2663 }
2664 }
2665 for (pI2o_dev = pHba->devices; pI2o_dev; pI2o_dev = pI2o_dev->next) {
2666 pDev =(struct adpt_device*) pI2o_dev->owner;
2667 if(!pDev){
2668 continue;
2669 }
2670 // Drive offline drives that previously existed but could not be found
2671 // in the LCT table
2672 if (pDev->state & DPTI_DEV_UNSCANNED){
2673 pDev->state = DPTI_DEV_OFFLINE;
9cb78c16 2674 printk(KERN_WARNING"%s: Device (%d,%d,%llu) offline\n",pHba->name,pDev->scsi_channel,pDev->scsi_id,pDev->scsi_lun);
1da177e4
LT
2675 if (pDev->pScsi_dev) {
2676 scsi_device_set_state(pDev->pScsi_dev, SDEV_OFFLINE);
2677 }
2678 }
2679 }
2680 return 0;
2681}
2682
2683static void adpt_fail_posted_scbs(adpt_hba* pHba)
2684{
2685 struct scsi_cmnd* cmd = NULL;
2686 struct scsi_device* d = NULL;
2687
2688 shost_for_each_device(d, pHba->host) {
2689 unsigned long flags;
2690 spin_lock_irqsave(&d->list_lock, flags);
2691 list_for_each_entry(cmd, &d->cmd_list, list) {
2692 if(cmd->serial_number == 0){
2693 continue;
2694 }
2695 cmd->result = (DID_OK << 16) | (QUEUE_FULL <<1);
2696 cmd->scsi_done(cmd);
2697 }
2698 spin_unlock_irqrestore(&d->list_lock, flags);
2699 }
2700}
2701
2702
2703/*============================================================================
2704 * Routines from i2o subsystem
2705 *============================================================================
2706 */
2707
2708
2709
2710/*
2711 * Bring an I2O controller into HOLD state. See the spec.
2712 */
2713static int adpt_i2o_activate_hba(adpt_hba* pHba)
2714{
2715 int rcode;
2716
2717 if(pHba->initialized ) {
2718 if (adpt_i2o_status_get(pHba) < 0) {
2719 if((rcode = adpt_i2o_reset_hba(pHba)) != 0){
2720 printk(KERN_WARNING"%s: Could NOT reset.\n", pHba->name);
2721 return rcode;
2722 }
2723 if (adpt_i2o_status_get(pHba) < 0) {
2724 printk(KERN_INFO "HBA not responding.\n");
2725 return -1;
2726 }
2727 }
2728
2729 if(pHba->status_block->iop_state == ADAPTER_STATE_FAULTED) {
2730 printk(KERN_CRIT "%s: hardware fault\n", pHba->name);
2731 return -1;
2732 }
2733
2734 if (pHba->status_block->iop_state == ADAPTER_STATE_READY ||
2735 pHba->status_block->iop_state == ADAPTER_STATE_OPERATIONAL ||
2736 pHba->status_block->iop_state == ADAPTER_STATE_HOLD ||
2737 pHba->status_block->iop_state == ADAPTER_STATE_FAILED) {
2738 adpt_i2o_reset_hba(pHba);
2739 if (adpt_i2o_status_get(pHba) < 0 || pHba->status_block->iop_state != ADAPTER_STATE_RESET) {
2740 printk(KERN_ERR "%s: Failed to initialize.\n", pHba->name);
2741 return -1;
2742 }
2743 }
2744 } else {
2745 if((rcode = adpt_i2o_reset_hba(pHba)) != 0){
2746 printk(KERN_WARNING"%s: Could NOT reset.\n", pHba->name);
2747 return rcode;
2748 }
2749
2750 }
2751
2752 if (adpt_i2o_init_outbound_q(pHba) < 0) {
2753 return -1;
2754 }
2755
2756 /* In HOLD state */
2757
2758 if (adpt_i2o_hrt_get(pHba) < 0) {
2759 return -1;
2760 }
2761
2762 return 0;
2763}
2764
2765/*
2766 * Bring a controller online into OPERATIONAL state.
2767 */
2768
2769static int adpt_i2o_online_hba(adpt_hba* pHba)
2770{
2771 if (adpt_i2o_systab_send(pHba) < 0) {
2772 adpt_i2o_delete_hba(pHba);
2773 return -1;
2774 }
2775 /* In READY state */
2776
2777 if (adpt_i2o_enable_hba(pHba) < 0) {
2778 adpt_i2o_delete_hba(pHba);
2779 return -1;
2780 }
2781
2782 /* In OPERATIONAL state */
2783 return 0;
2784}
2785
2786static s32 adpt_send_nop(adpt_hba*pHba,u32 m)
2787{
2788 u32 __iomem *msg;
2789 ulong timeout = jiffies + 5*HZ;
2790
2791 while(m == EMPTY_QUEUE){
2792 rmb();
2793 m = readl(pHba->post_port);
2794 if(m != EMPTY_QUEUE){
2795 break;
2796 }
2797 if(time_after(jiffies,timeout)){
2798 printk(KERN_ERR "%s: Timeout waiting for message frame!\n",pHba->name);
2799 return 2;
2800 }
a9a3047d 2801 schedule_timeout_uninterruptible(1);
1da177e4
LT
2802 }
2803 msg = (u32 __iomem *)(pHba->msg_addr_virt + m);
2804 writel( THREE_WORD_MSG_SIZE | SGL_OFFSET_0,&msg[0]);
2805 writel( I2O_CMD_UTIL_NOP << 24 | HOST_TID << 12 | 0,&msg[1]);
2806 writel( 0,&msg[2]);
2807 wmb();
2808
2809 writel(m, pHba->post_port);
2810 wmb();
2811 return 0;
2812}
2813
2814static s32 adpt_i2o_init_outbound_q(adpt_hba* pHba)
2815{
2816 u8 *status;
67af2b06 2817 dma_addr_t addr;
1da177e4
LT
2818 u32 __iomem *msg = NULL;
2819 int i;
2820 ulong timeout = jiffies + TMOUT_INITOUTBOUND*HZ;
1da177e4
LT
2821 u32 m;
2822
2823 do {
2824 rmb();
2825 m = readl(pHba->post_port);
2826 if (m != EMPTY_QUEUE) {
2827 break;
2828 }
2829
2830 if(time_after(jiffies,timeout)){
2831 printk(KERN_WARNING"%s: Timeout waiting for message frame\n",pHba->name);
2832 return -ETIMEDOUT;
2833 }
a9a3047d 2834 schedule_timeout_uninterruptible(1);
1da177e4
LT
2835 } while(m == EMPTY_QUEUE);
2836
2837 msg=(u32 __iomem *)(pHba->msg_addr_virt+m);
2838
67af2b06 2839 status = dma_alloc_coherent(&pHba->pDev->dev, 4, &addr, GFP_KERNEL);
bbfbbbc1 2840 if (!status) {
1da177e4
LT
2841 adpt_send_nop(pHba, m);
2842 printk(KERN_WARNING"%s: IOP reset failed - no free memory.\n",
2843 pHba->name);
2844 return -ENOMEM;
2845 }
67af2b06 2846 memset(status, 0, 4);
1da177e4
LT
2847
2848 writel(EIGHT_WORD_MSG_SIZE| SGL_OFFSET_6, &msg[0]);
2849 writel(I2O_CMD_OUTBOUND_INIT<<24 | HOST_TID<<12 | ADAPTER_TID, &msg[1]);
2850 writel(0, &msg[2]);
2851 writel(0x0106, &msg[3]); /* Transaction context */
2852 writel(4096, &msg[4]); /* Host page frame size */
2853 writel((REPLY_FRAME_SIZE)<<16|0x80, &msg[5]); /* Outbound msg frame size and Initcode */
2854 writel(0xD0000004, &msg[6]); /* Simple SG LE, EOB */
67af2b06 2855 writel((u32)addr, &msg[7]);
1da177e4
LT
2856
2857 writel(m, pHba->post_port);
2858 wmb();
2859
2860 // Wait for the reply status to come back
2861 do {
2862 if (*status) {
2863 if (*status != 0x01 /*I2O_EXEC_OUTBOUND_INIT_IN_PROGRESS*/) {
2864 break;
2865 }
2866 }
2867 rmb();
2868 if(time_after(jiffies,timeout)){
2869 printk(KERN_WARNING"%s: Timeout Initializing\n",pHba->name);
67af2b06
MS
2870 /* We lose 4 bytes of "status" here, but we
2871 cannot free these because controller may
2872 awake and corrupt those bytes at any time */
2873 /* dma_free_coherent(&pHba->pDev->dev, 4, status, addr); */
1da177e4
LT
2874 return -ETIMEDOUT;
2875 }
a9a3047d 2876 schedule_timeout_uninterruptible(1);
1da177e4
LT
2877 } while (1);
2878
2879 // If the command was successful, fill the fifo with our reply
2880 // message packets
2881 if(*status != 0x04 /*I2O_EXEC_OUTBOUND_INIT_COMPLETE*/) {
67af2b06 2882 dma_free_coherent(&pHba->pDev->dev, 4, status, addr);
1da177e4
LT
2883 return -2;
2884 }
67af2b06 2885 dma_free_coherent(&pHba->pDev->dev, 4, status, addr);
1da177e4 2886
67af2b06
MS
2887 if(pHba->reply_pool != NULL) {
2888 dma_free_coherent(&pHba->pDev->dev,
2889 pHba->reply_fifo_size * REPLY_FRAME_SIZE * 4,
2890 pHba->reply_pool, pHba->reply_pool_pa);
2891 }
1da177e4 2892
67af2b06
MS
2893 pHba->reply_pool = dma_alloc_coherent(&pHba->pDev->dev,
2894 pHba->reply_fifo_size * REPLY_FRAME_SIZE * 4,
2895 &pHba->reply_pool_pa, GFP_KERNEL);
bbfbbbc1
MK
2896 if (!pHba->reply_pool) {
2897 printk(KERN_ERR "%s: Could not allocate reply pool\n", pHba->name);
2898 return -ENOMEM;
1da177e4 2899 }
67af2b06 2900 memset(pHba->reply_pool, 0 , pHba->reply_fifo_size * REPLY_FRAME_SIZE * 4);
1da177e4 2901
1da177e4 2902 for(i = 0; i < pHba->reply_fifo_size; i++) {
67af2b06
MS
2903 writel(pHba->reply_pool_pa + (i * REPLY_FRAME_SIZE * 4),
2904 pHba->reply_port);
1da177e4 2905 wmb();
1da177e4
LT
2906 }
2907 adpt_i2o_status_get(pHba);
2908 return 0;
2909}
2910
2911
2912/*
2913 * I2O System Table. Contains information about
2914 * all the IOPs in the system. Used to inform IOPs
2915 * about each other's existence.
2916 *
2917 * sys_tbl_ver is the CurrentChangeIndicator that is
2918 * used by IOPs to track changes.
2919 */
2920
2921
2922
2923static s32 adpt_i2o_status_get(adpt_hba* pHba)
2924{
2925 ulong timeout;
2926 u32 m;
2927 u32 __iomem *msg;
2928 u8 *status_block=NULL;
1da177e4
LT
2929
2930 if(pHba->status_block == NULL) {
67af2b06
MS
2931 pHba->status_block = dma_alloc_coherent(&pHba->pDev->dev,
2932 sizeof(i2o_status_block),
2933 &pHba->status_block_pa, GFP_KERNEL);
1da177e4
LT
2934 if(pHba->status_block == NULL) {
2935 printk(KERN_ERR
2936 "dpti%d: Get Status Block failed; Out of memory. \n",
2937 pHba->unit);
2938 return -ENOMEM;
2939 }
2940 }
2941 memset(pHba->status_block, 0, sizeof(i2o_status_block));
2942 status_block = (u8*)(pHba->status_block);
1da177e4
LT
2943 timeout = jiffies+TMOUT_GETSTATUS*HZ;
2944 do {
2945 rmb();
2946 m = readl(pHba->post_port);
2947 if (m != EMPTY_QUEUE) {
2948 break;
2949 }
2950 if(time_after(jiffies,timeout)){
2951 printk(KERN_ERR "%s: Timeout waiting for message !\n",
2952 pHba->name);
2953 return -ETIMEDOUT;
2954 }
a9a3047d 2955 schedule_timeout_uninterruptible(1);
1da177e4
LT
2956 } while(m==EMPTY_QUEUE);
2957
2958
2959 msg=(u32 __iomem *)(pHba->msg_addr_virt+m);
2960
2961 writel(NINE_WORD_MSG_SIZE|SGL_OFFSET_0, &msg[0]);
2962 writel(I2O_CMD_STATUS_GET<<24|HOST_TID<<12|ADAPTER_TID, &msg[1]);
2963 writel(1, &msg[2]);
2964 writel(0, &msg[3]);
2965 writel(0, &msg[4]);
2966 writel(0, &msg[5]);
67af2b06
MS
2967 writel( dma_low(pHba->status_block_pa), &msg[6]);
2968 writel( dma_high(pHba->status_block_pa), &msg[7]);
1da177e4
LT
2969 writel(sizeof(i2o_status_block), &msg[8]); // 88 bytes
2970
2971 //post message
2972 writel(m, pHba->post_port);
2973 wmb();
2974
2975 while(status_block[87]!=0xff){
2976 if(time_after(jiffies,timeout)){
2977 printk(KERN_ERR"dpti%d: Get status timeout.\n",
2978 pHba->unit);
2979 return -ETIMEDOUT;
2980 }
2981 rmb();
a9a3047d 2982 schedule_timeout_uninterruptible(1);
1da177e4
LT
2983 }
2984
2985 // Set up our number of outbound and inbound messages
2986 pHba->post_fifo_size = pHba->status_block->max_inbound_frames;
2987 if (pHba->post_fifo_size > MAX_TO_IOP_MESSAGES) {
2988 pHba->post_fifo_size = MAX_TO_IOP_MESSAGES;
2989 }
2990
2991 pHba->reply_fifo_size = pHba->status_block->max_outbound_frames;
2992 if (pHba->reply_fifo_size > MAX_FROM_IOP_MESSAGES) {
2993 pHba->reply_fifo_size = MAX_FROM_IOP_MESSAGES;
2994 }
2995
2996 // Calculate the Scatter Gather list size
62ac5aed
MS
2997 if (dpt_dma64(pHba)) {
2998 pHba->sg_tablesize
2999 = ((pHba->status_block->inbound_frame_size * 4
3000 - 14 * sizeof(u32))
3001 / (sizeof(struct sg_simple_element) + sizeof(u32)));
3002 } else {
3003 pHba->sg_tablesize
3004 = ((pHba->status_block->inbound_frame_size * 4
3005 - 12 * sizeof(u32))
3006 / sizeof(struct sg_simple_element));
3007 }
1da177e4
LT
3008 if (pHba->sg_tablesize > SG_LIST_ELEMENTS) {
3009 pHba->sg_tablesize = SG_LIST_ELEMENTS;
3010 }
3011
3012
3013#ifdef DEBUG
3014 printk("dpti%d: State = ",pHba->unit);
3015 switch(pHba->status_block->iop_state) {
3016 case 0x01:
3017 printk("INIT\n");
3018 break;
3019 case 0x02:
3020 printk("RESET\n");
3021 break;
3022 case 0x04:
3023 printk("HOLD\n");
3024 break;
3025 case 0x05:
3026 printk("READY\n");
3027 break;
3028 case 0x08:
3029 printk("OPERATIONAL\n");
3030 break;
3031 case 0x10:
3032 printk("FAILED\n");
3033 break;
3034 case 0x11:
3035 printk("FAULTED\n");
3036 break;
3037 default:
3038 printk("%x (unknown!!)\n",pHba->status_block->iop_state);
3039 }
3040#endif
3041 return 0;
3042}
3043
3044/*
3045 * Get the IOP's Logical Configuration Table
3046 */
3047static int adpt_i2o_lct_get(adpt_hba* pHba)
3048{
3049 u32 msg[8];
3050 int ret;
3051 u32 buf[16];
3052
3053 if ((pHba->lct_size == 0) || (pHba->lct == NULL)){
3054 pHba->lct_size = pHba->status_block->expected_lct_size;
3055 }
3056 do {
3057 if (pHba->lct == NULL) {
67af2b06
MS
3058 pHba->lct = dma_alloc_coherent(&pHba->pDev->dev,
3059 pHba->lct_size, &pHba->lct_pa,
da2907ff 3060 GFP_ATOMIC);
1da177e4
LT
3061 if(pHba->lct == NULL) {
3062 printk(KERN_CRIT "%s: Lct Get failed. Out of memory.\n",
3063 pHba->name);
3064 return -ENOMEM;
3065 }
3066 }
3067 memset(pHba->lct, 0, pHba->lct_size);
3068
3069 msg[0] = EIGHT_WORD_MSG_SIZE|SGL_OFFSET_6;
3070 msg[1] = I2O_CMD_LCT_NOTIFY<<24 | HOST_TID<<12 | ADAPTER_TID;
3071 msg[2] = 0;
3072 msg[3] = 0;
3073 msg[4] = 0xFFFFFFFF; /* All devices */
3074 msg[5] = 0x00000000; /* Report now */
3075 msg[6] = 0xD0000000|pHba->lct_size;
67af2b06 3076 msg[7] = (u32)pHba->lct_pa;
1da177e4
LT
3077
3078 if ((ret=adpt_i2o_post_wait(pHba, msg, sizeof(msg), 360))) {
3079 printk(KERN_ERR "%s: LCT Get failed (status=%#10x.\n",
3080 pHba->name, ret);
3081 printk(KERN_ERR"Adaptec: Error Reading Hardware.\n");
3082 return ret;
3083 }
3084
3085 if ((pHba->lct->table_size << 2) > pHba->lct_size) {
3086 pHba->lct_size = pHba->lct->table_size << 2;
67af2b06
MS
3087 dma_free_coherent(&pHba->pDev->dev, pHba->lct_size,
3088 pHba->lct, pHba->lct_pa);
1da177e4
LT
3089 pHba->lct = NULL;
3090 }
3091 } while (pHba->lct == NULL);
3092
3093 PDEBUG("%s: Hardware resource table read.\n", pHba->name);
3094
3095
3096 // I2O_DPT_EXEC_IOP_BUFFERS_GROUP_NO;
3097 if(adpt_i2o_query_scalar(pHba, 0 , 0x8000, -1, buf, sizeof(buf))>=0) {
3098 pHba->FwDebugBufferSize = buf[1];
62ac5aed
MS
3099 pHba->FwDebugBuffer_P = ioremap(pHba->base_addr_phys + buf[0],
3100 pHba->FwDebugBufferSize);
3101 if (pHba->FwDebugBuffer_P) {
3102 pHba->FwDebugFlags_P = pHba->FwDebugBuffer_P +
3103 FW_DEBUG_FLAGS_OFFSET;
3104 pHba->FwDebugBLEDvalue_P = pHba->FwDebugBuffer_P +
3105 FW_DEBUG_BLED_OFFSET;
3106 pHba->FwDebugBLEDflag_P = pHba->FwDebugBLEDvalue_P + 1;
3107 pHba->FwDebugStrLength_P = pHba->FwDebugBuffer_P +
3108 FW_DEBUG_STR_LENGTH_OFFSET;
3109 pHba->FwDebugBuffer_P += buf[2];
3110 pHba->FwDebugFlags = 0;
3111 }
1da177e4
LT
3112 }
3113
3114 return 0;
3115}
3116
3117static int adpt_i2o_build_sys_table(void)
3118{
67af2b06 3119 adpt_hba* pHba = hba_chain;
1da177e4
LT
3120 int count = 0;
3121
67af2b06
MS
3122 if (sys_tbl)
3123 dma_free_coherent(&pHba->pDev->dev, sys_tbl_len,
3124 sys_tbl, sys_tbl_pa);
3125
1da177e4
LT
3126 sys_tbl_len = sizeof(struct i2o_sys_tbl) + // Header + IOPs
3127 (hba_count) * sizeof(struct i2o_sys_tbl_entry);
3128
67af2b06
MS
3129 sys_tbl = dma_alloc_coherent(&pHba->pDev->dev,
3130 sys_tbl_len, &sys_tbl_pa, GFP_KERNEL);
bbfbbbc1 3131 if (!sys_tbl) {
1da177e4
LT
3132 printk(KERN_WARNING "SysTab Set failed. Out of memory.\n");
3133 return -ENOMEM;
3134 }
67af2b06 3135 memset(sys_tbl, 0, sys_tbl_len);
1da177e4
LT
3136
3137 sys_tbl->num_entries = hba_count;
3138 sys_tbl->version = I2OVERSION;
3139 sys_tbl->change_ind = sys_tbl_ind++;
3140
3141 for(pHba = hba_chain; pHba; pHba = pHba->next) {
67af2b06 3142 u64 addr;
1da177e4
LT
3143 // Get updated Status Block so we have the latest information
3144 if (adpt_i2o_status_get(pHba)) {
3145 sys_tbl->num_entries--;
3146 continue; // try next one
3147 }
3148
3149 sys_tbl->iops[count].org_id = pHba->status_block->org_id;
3150 sys_tbl->iops[count].iop_id = pHba->unit + 2;
3151 sys_tbl->iops[count].seg_num = 0;
3152 sys_tbl->iops[count].i2o_version = pHba->status_block->i2o_version;
3153 sys_tbl->iops[count].iop_state = pHba->status_block->iop_state;
3154 sys_tbl->iops[count].msg_type = pHba->status_block->msg_type;
3155 sys_tbl->iops[count].frame_size = pHba->status_block->inbound_frame_size;
3156 sys_tbl->iops[count].last_changed = sys_tbl_ind - 1; // ??
3157 sys_tbl->iops[count].iop_capabilities = pHba->status_block->iop_capabilities;
67af2b06
MS
3158 addr = pHba->base_addr_phys + 0x40;
3159 sys_tbl->iops[count].inbound_low = dma_low(addr);
3160 sys_tbl->iops[count].inbound_high = dma_high(addr);
1da177e4
LT
3161
3162 count++;
3163 }
3164
3165#ifdef DEBUG
3166{
3167 u32 *table = (u32*)sys_tbl;
3168 printk(KERN_DEBUG"sys_tbl_len=%d in 32bit words\n",(sys_tbl_len >>2));
3169 for(count = 0; count < (sys_tbl_len >>2); count++) {
3170 printk(KERN_INFO "sys_tbl[%d] = %0#10x\n",
3171 count, table[count]);
3172 }
3173}
3174#endif
3175
3176 return 0;
3177}
3178
3179
3180/*
3181 * Dump the information block associated with a given unit (TID)
3182 */
3183
3184static void adpt_i2o_report_hba_unit(adpt_hba* pHba, struct i2o_device *d)
3185{
3186 char buf[64];
3187 int unit = d->lct_data.tid;
3188
3189 printk(KERN_INFO "TID %3.3d ", unit);
3190
3191 if(adpt_i2o_query_scalar(pHba, unit, 0xF100, 3, buf, 16)>=0)
3192 {
3193 buf[16]=0;
3194 printk(" Vendor: %-12.12s", buf);
3195 }
3196 if(adpt_i2o_query_scalar(pHba, unit, 0xF100, 4, buf, 16)>=0)
3197 {
3198 buf[16]=0;
3199 printk(" Device: %-12.12s", buf);
3200 }
3201 if(adpt_i2o_query_scalar(pHba, unit, 0xF100, 6, buf, 8)>=0)
3202 {
3203 buf[8]=0;
3204 printk(" Rev: %-12.12s\n", buf);
3205 }
3206#ifdef DEBUG
3207 printk(KERN_INFO "\tClass: %.21s\n", adpt_i2o_get_class_name(d->lct_data.class_id));
3208 printk(KERN_INFO "\tSubclass: 0x%04X\n", d->lct_data.sub_class);
3209 printk(KERN_INFO "\tFlags: ");
3210
3211 if(d->lct_data.device_flags&(1<<0))
3212 printk("C"); // ConfigDialog requested
3213 if(d->lct_data.device_flags&(1<<1))
3214 printk("U"); // Multi-user capable
3215 if(!(d->lct_data.device_flags&(1<<4)))
3216 printk("P"); // Peer service enabled!
3217 if(!(d->lct_data.device_flags&(1<<5)))
3218 printk("M"); // Mgmt service enabled!
3219 printk("\n");
3220#endif
3221}
3222
3223#ifdef DEBUG
3224/*
3225 * Do i2o class name lookup
3226 */
3227static const char *adpt_i2o_get_class_name(int class)
3228{
3229 int idx = 16;
3230 static char *i2o_class_name[] = {
3231 "Executive",
3232 "Device Driver Module",
3233 "Block Device",
3234 "Tape Device",
3235 "LAN Interface",
3236 "WAN Interface",
3237 "Fibre Channel Port",
3238 "Fibre Channel Device",
3239 "SCSI Device",
3240 "ATE Port",
3241 "ATE Device",
3242 "Floppy Controller",
3243 "Floppy Device",
3244 "Secondary Bus Port",
3245 "Peer Transport Agent",
3246 "Peer Transport",
3247 "Unknown"
3248 };
3249
3250 switch(class&0xFFF) {
3251 case I2O_CLASS_EXECUTIVE:
3252 idx = 0; break;
3253 case I2O_CLASS_DDM:
3254 idx = 1; break;
3255 case I2O_CLASS_RANDOM_BLOCK_STORAGE:
3256 idx = 2; break;
3257 case I2O_CLASS_SEQUENTIAL_STORAGE:
3258 idx = 3; break;
3259 case I2O_CLASS_LAN:
3260 idx = 4; break;
3261 case I2O_CLASS_WAN:
3262 idx = 5; break;
3263 case I2O_CLASS_FIBRE_CHANNEL_PORT:
3264 idx = 6; break;
3265 case I2O_CLASS_FIBRE_CHANNEL_PERIPHERAL:
3266 idx = 7; break;
3267 case I2O_CLASS_SCSI_PERIPHERAL:
3268 idx = 8; break;
3269 case I2O_CLASS_ATE_PORT:
3270 idx = 9; break;
3271 case I2O_CLASS_ATE_PERIPHERAL:
3272 idx = 10; break;
3273 case I2O_CLASS_FLOPPY_CONTROLLER:
3274 idx = 11; break;
3275 case I2O_CLASS_FLOPPY_DEVICE:
3276 idx = 12; break;
3277 case I2O_CLASS_BUS_ADAPTER_PORT:
3278 idx = 13; break;
3279 case I2O_CLASS_PEER_TRANSPORT_AGENT:
3280 idx = 14; break;
3281 case I2O_CLASS_PEER_TRANSPORT:
3282 idx = 15; break;
3283 }
3284 return i2o_class_name[idx];
3285}
3286#endif
3287
3288
3289static s32 adpt_i2o_hrt_get(adpt_hba* pHba)
3290{
3291 u32 msg[6];
3292 int ret, size = sizeof(i2o_hrt);
3293
3294 do {
3295 if (pHba->hrt == NULL) {
67af2b06
MS
3296 pHba->hrt = dma_alloc_coherent(&pHba->pDev->dev,
3297 size, &pHba->hrt_pa, GFP_KERNEL);
1da177e4
LT
3298 if (pHba->hrt == NULL) {
3299 printk(KERN_CRIT "%s: Hrt Get failed; Out of memory.\n", pHba->name);
3300 return -ENOMEM;
3301 }
3302 }
3303
3304 msg[0]= SIX_WORD_MSG_SIZE| SGL_OFFSET_4;
3305 msg[1]= I2O_CMD_HRT_GET<<24 | HOST_TID<<12 | ADAPTER_TID;
3306 msg[2]= 0;
3307 msg[3]= 0;
3308 msg[4]= (0xD0000000 | size); /* Simple transaction */
67af2b06 3309 msg[5]= (u32)pHba->hrt_pa; /* Dump it here */
1da177e4
LT
3310
3311 if ((ret = adpt_i2o_post_wait(pHba, msg, sizeof(msg),20))) {
3312 printk(KERN_ERR "%s: Unable to get HRT (status=%#10x)\n", pHba->name, ret);
3313 return ret;
3314 }
3315
3316 if (pHba->hrt->num_entries * pHba->hrt->entry_len << 2 > size) {
67af2b06
MS
3317 int newsize = pHba->hrt->num_entries * pHba->hrt->entry_len << 2;
3318 dma_free_coherent(&pHba->pDev->dev, size,
3319 pHba->hrt, pHba->hrt_pa);
3320 size = newsize;
1da177e4
LT
3321 pHba->hrt = NULL;
3322 }
3323 } while(pHba->hrt == NULL);
3324 return 0;
3325}
3326
3327/*
3328 * Query one scalar group value or a whole scalar group.
3329 */
3330static int adpt_i2o_query_scalar(adpt_hba* pHba, int tid,
3331 int group, int field, void *buf, int buflen)
3332{
3333 u16 opblk[] = { 1, 0, I2O_PARAMS_FIELD_GET, group, 1, field };
67af2b06
MS
3334 u8 *opblk_va;
3335 dma_addr_t opblk_pa;
3336 u8 *resblk_va;
3337 dma_addr_t resblk_pa;
1da177e4
LT
3338
3339 int size;
3340
3341 /* 8 bytes for header */
67af2b06
MS
3342 resblk_va = dma_alloc_coherent(&pHba->pDev->dev,
3343 sizeof(u8) * (8 + buflen), &resblk_pa, GFP_KERNEL);
3344 if (resblk_va == NULL) {
1da177e4
LT
3345 printk(KERN_CRIT "%s: query scalar failed; Out of memory.\n", pHba->name);
3346 return -ENOMEM;
3347 }
3348
67af2b06
MS
3349 opblk_va = dma_alloc_coherent(&pHba->pDev->dev,
3350 sizeof(opblk), &opblk_pa, GFP_KERNEL);
3351 if (opblk_va == NULL) {
3352 dma_free_coherent(&pHba->pDev->dev, sizeof(u8) * (8+buflen),
3353 resblk_va, resblk_pa);
d0799621 3354 printk(KERN_CRIT "%s: query operation failed; Out of memory.\n",
67af2b06
MS
3355 pHba->name);
3356 return -ENOMEM;
3357 }
1da177e4
LT
3358 if (field == -1) /* whole group */
3359 opblk[4] = -1;
3360
67af2b06 3361 memcpy(opblk_va, opblk, sizeof(opblk));
1da177e4 3362 size = adpt_i2o_issue_params(I2O_CMD_UTIL_PARAMS_GET, pHba, tid,
67af2b06
MS
3363 opblk_va, opblk_pa, sizeof(opblk),
3364 resblk_va, resblk_pa, sizeof(u8)*(8+buflen));
3365 dma_free_coherent(&pHba->pDev->dev, sizeof(opblk), opblk_va, opblk_pa);
1da177e4 3366 if (size == -ETIME) {
67af2b06
MS
3367 dma_free_coherent(&pHba->pDev->dev, sizeof(u8) * (8+buflen),
3368 resblk_va, resblk_pa);
1da177e4
LT
3369 printk(KERN_WARNING "%s: issue params failed; Timed out.\n", pHba->name);
3370 return -ETIME;
3371 } else if (size == -EINTR) {
67af2b06
MS
3372 dma_free_coherent(&pHba->pDev->dev, sizeof(u8) * (8+buflen),
3373 resblk_va, resblk_pa);
1da177e4
LT
3374 printk(KERN_WARNING "%s: issue params failed; Interrupted.\n", pHba->name);
3375 return -EINTR;
3376 }
3377
67af2b06 3378 memcpy(buf, resblk_va+8, buflen); /* cut off header */
1da177e4 3379
67af2b06
MS
3380 dma_free_coherent(&pHba->pDev->dev, sizeof(u8) * (8+buflen),
3381 resblk_va, resblk_pa);
1da177e4
LT
3382 if (size < 0)
3383 return size;
3384
3385 return buflen;
3386}
3387
3388
3389/* Issue UTIL_PARAMS_GET or UTIL_PARAMS_SET
3390 *
3391 * This function can be used for all UtilParamsGet/Set operations.
3392 * The OperationBlock is given in opblk-buffer,
3393 * and results are returned in resblk-buffer.
3394 * Note that the minimum sized resblk is 8 bytes and contains
3395 * ResultCount, ErrorInfoSize, BlockStatus and BlockSize.
3396 */
3397static int adpt_i2o_issue_params(int cmd, adpt_hba* pHba, int tid,
67af2b06
MS
3398 void *opblk_va, dma_addr_t opblk_pa, int oplen,
3399 void *resblk_va, dma_addr_t resblk_pa, int reslen)
1da177e4
LT
3400{
3401 u32 msg[9];
67af2b06 3402 u32 *res = (u32 *)resblk_va;
1da177e4
LT
3403 int wait_status;
3404
3405 msg[0] = NINE_WORD_MSG_SIZE | SGL_OFFSET_5;
3406 msg[1] = cmd << 24 | HOST_TID << 12 | tid;
3407 msg[2] = 0;
3408 msg[3] = 0;
3409 msg[4] = 0;
3410 msg[5] = 0x54000000 | oplen; /* OperationBlock */
67af2b06 3411 msg[6] = (u32)opblk_pa;
1da177e4 3412 msg[7] = 0xD0000000 | reslen; /* ResultBlock */
67af2b06 3413 msg[8] = (u32)resblk_pa;
1da177e4
LT
3414
3415 if ((wait_status = adpt_i2o_post_wait(pHba, msg, sizeof(msg), 20))) {
67af2b06 3416 printk("adpt_i2o_issue_params: post_wait failed (%p)\n", resblk_va);
1da177e4
LT
3417 return wait_status; /* -DetailedStatus */
3418 }
3419
3420 if (res[1]&0x00FF0000) { /* BlockStatus != SUCCESS */
3421 printk(KERN_WARNING "%s: %s - Error:\n ErrorInfoSize = 0x%02x, "
3422 "BlockStatus = 0x%02x, BlockSize = 0x%04x\n",
3423 pHba->name,
3424 (cmd == I2O_CMD_UTIL_PARAMS_SET) ? "PARAMS_SET"
3425 : "PARAMS_GET",
3426 res[1]>>24, (res[1]>>16)&0xFF, res[1]&0xFFFF);
3427 return -((res[1] >> 16) & 0xFF); /* -BlockStatus */
3428 }
3429
3430 return 4 + ((res[1] & 0x0000FFFF) << 2); /* bytes used in resblk */
3431}
3432
3433
3434static s32 adpt_i2o_quiesce_hba(adpt_hba* pHba)
3435{
3436 u32 msg[4];
3437 int ret;
3438
3439 adpt_i2o_status_get(pHba);
3440
3441 /* SysQuiesce discarded if IOP not in READY or OPERATIONAL state */
3442
3443 if((pHba->status_block->iop_state != ADAPTER_STATE_READY) &&
3444 (pHba->status_block->iop_state != ADAPTER_STATE_OPERATIONAL)){
3445 return 0;
3446 }
3447
3448 msg[0] = FOUR_WORD_MSG_SIZE|SGL_OFFSET_0;
3449 msg[1] = I2O_CMD_SYS_QUIESCE<<24|HOST_TID<<12|ADAPTER_TID;
3450 msg[2] = 0;
3451 msg[3] = 0;
3452
3453 if((ret = adpt_i2o_post_wait(pHba, msg, sizeof(msg), 240))) {
3454 printk(KERN_INFO"dpti%d: Unable to quiesce (status=%#x).\n",
3455 pHba->unit, -ret);
3456 } else {
3457 printk(KERN_INFO"dpti%d: Quiesced.\n",pHba->unit);
3458 }
3459
3460 adpt_i2o_status_get(pHba);
3461 return ret;
3462}
3463
3464
3465/*
3466 * Enable IOP. Allows the IOP to resume external operations.
3467 */
3468static int adpt_i2o_enable_hba(adpt_hba* pHba)
3469{
3470 u32 msg[4];
3471 int ret;
3472
3473 adpt_i2o_status_get(pHba);
3474 if(!pHba->status_block){
3475 return -ENOMEM;
3476 }
3477 /* Enable only allowed on READY state */
3478 if(pHba->status_block->iop_state == ADAPTER_STATE_OPERATIONAL)
3479 return 0;
3480
3481 if(pHba->status_block->iop_state != ADAPTER_STATE_READY)
3482 return -EINVAL;
3483
3484 msg[0]=FOUR_WORD_MSG_SIZE|SGL_OFFSET_0;
3485 msg[1]=I2O_CMD_SYS_ENABLE<<24|HOST_TID<<12|ADAPTER_TID;
3486 msg[2]= 0;
3487 msg[3]= 0;
3488
3489 if ((ret = adpt_i2o_post_wait(pHba, msg, sizeof(msg), 240))) {
3490 printk(KERN_WARNING"%s: Could not enable (status=%#10x).\n",
3491 pHba->name, ret);
3492 } else {
3493 PDEBUG("%s: Enabled.\n", pHba->name);
3494 }
3495
3496 adpt_i2o_status_get(pHba);
3497 return ret;
3498}
3499
3500
3501static int adpt_i2o_systab_send(adpt_hba* pHba)
3502{
3503 u32 msg[12];
3504 int ret;
3505
3506 msg[0] = I2O_MESSAGE_SIZE(12) | SGL_OFFSET_6;
3507 msg[1] = I2O_CMD_SYS_TAB_SET<<24 | HOST_TID<<12 | ADAPTER_TID;
3508 msg[2] = 0;
3509 msg[3] = 0;
3510 msg[4] = (0<<16) | ((pHba->unit+2) << 12); /* Host 0 IOP ID (unit + 2) */
3511 msg[5] = 0; /* Segment 0 */
3512
3513 /*
3514 * Provide three SGL-elements:
3515 * System table (SysTab), Private memory space declaration and
3516 * Private i/o space declaration
3517 */
3518 msg[6] = 0x54000000 | sys_tbl_len;
67af2b06 3519 msg[7] = (u32)sys_tbl_pa;
1da177e4
LT
3520 msg[8] = 0x54000000 | 0;
3521 msg[9] = 0;
3522 msg[10] = 0xD4000000 | 0;
3523 msg[11] = 0;
3524
3525 if ((ret=adpt_i2o_post_wait(pHba, msg, sizeof(msg), 120))) {
3526 printk(KERN_INFO "%s: Unable to set SysTab (status=%#10x).\n",
3527 pHba->name, ret);
3528 }
3529#ifdef DEBUG
3530 else {
3531 PINFO("%s: SysTab set.\n", pHba->name);
3532 }
3533#endif
3534
3535 return ret;
3536 }
3537
3538
3539/*============================================================================
3540 *
3541 *============================================================================
3542 */
3543
3544
3545#ifdef UARTDELAY
3546
3547static static void adpt_delay(int millisec)
3548{
3549 int i;
3550 for (i = 0; i < millisec; i++) {
3551 udelay(1000); /* delay for one millisecond */
3552 }
3553}
3554
3555#endif
3556
24601bbc 3557static struct scsi_host_template driver_template = {
c864cb14 3558 .module = THIS_MODULE,
1da177e4
LT
3559 .name = "dpt_i2o",
3560 .proc_name = "dpt_i2o",
ff98f7ce 3561 .show_info = adpt_show_info,
1da177e4
LT
3562 .info = adpt_info,
3563 .queuecommand = adpt_queue,
3564 .eh_abort_handler = adpt_abort,
3565 .eh_device_reset_handler = adpt_device_reset,
3566 .eh_bus_reset_handler = adpt_bus_reset,
3567 .eh_host_reset_handler = adpt_reset,
3568 .bios_param = adpt_bios_param,
3569 .slave_configure = adpt_slave_configure,
3570 .can_queue = MAX_TO_IOP_MESSAGES,
3571 .this_id = 7,
1da177e4
LT
3572 .use_clustering = ENABLE_CLUSTERING,
3573};
c864cb14
MS
3574
3575static int __init adpt_init(void)
3576{
3577 int error;
3578 adpt_hba *pHba, *next;
3579
3580 printk("Loading Adaptec I2O RAID: Version " DPT_I2O_VERSION "\n");
3581
3582 error = adpt_detect(&driver_template);
3583 if (error < 0)
3584 return error;
3585 if (hba_chain == NULL)
3586 return -ENODEV;
3587
3588 for (pHba = hba_chain; pHba; pHba = pHba->next) {
3589 error = scsi_add_host(pHba->host, &pHba->pDev->dev);
3590 if (error)
3591 goto fail;
3592 scsi_scan_host(pHba->host);
3593 }
3594 return 0;
3595fail:
3596 for (pHba = hba_chain; pHba; pHba = next) {
3597 next = pHba->next;
3598 scsi_remove_host(pHba->host);
3599 }
3600 return error;
3601}
3602
3603static void __exit adpt_exit(void)
3604{
3605 adpt_hba *pHba, *next;
3606
3607 for (pHba = hba_chain; pHba; pHba = pHba->next)
3608 scsi_remove_host(pHba->host);
3609 for (pHba = hba_chain; pHba; pHba = next) {
3610 next = pHba->next;
3611 adpt_release(pHba->host);
3612 }
3613}
3614
3615module_init(adpt_init);
3616module_exit(adpt_exit);
3617
1da177e4 3618MODULE_LICENSE("GPL");