]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - drivers/scsi/53c700.c
Merge tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm
[mirror_ubuntu-artful-kernel.git] / drivers / scsi / 53c700.c
1 /* -*- mode: c; c-basic-offset: 8 -*- */
2
3 /* NCR (or Symbios) 53c700 and 53c700-66 Driver
4 *
5 * Copyright (C) 2001 by James.Bottomley@HansenPartnership.com
6 **-----------------------------------------------------------------------------
7 **
8 ** This program is free software; you can redistribute it and/or modify
9 ** it under the terms of the GNU General Public License as published by
10 ** the Free Software Foundation; either version 2 of the License, or
11 ** (at your option) any later version.
12 **
13 ** This program is distributed in the hope that it will be useful,
14 ** but WITHOUT ANY WARRANTY; without even the implied warranty of
15 ** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 ** GNU General Public License for more details.
17 **
18 ** You should have received a copy of the GNU General Public License
19 ** along with this program; if not, write to the Free Software
20 ** Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
21 **
22 **-----------------------------------------------------------------------------
23 */
24
25 /* Notes:
26 *
27 * This driver is designed exclusively for these chips (virtually the
28 * earliest of the scripts engine chips). They need their own drivers
29 * because they are missing so many of the scripts and snazzy register
30 * features of their elder brothers (the 710, 720 and 770).
31 *
32 * The 700 is the lowliest of the line, it can only do async SCSI.
33 * The 700-66 can at least do synchronous SCSI up to 10MHz.
34 *
35 * The 700 chip has no host bus interface logic of its own. However,
36 * it is usually mapped to a location with well defined register
37 * offsets. Therefore, if you can determine the base address and the
38 * irq your board incorporating this chip uses, you can probably use
39 * this driver to run it (although you'll probably have to write a
40 * minimal wrapper for the purpose---see the NCR_D700 driver for
41 * details about how to do this).
42 *
43 *
44 * TODO List:
45 *
46 * 1. Better statistics in the proc fs
47 *
48 * 2. Implement message queue (queues SCSI messages like commands) and make
49 * the abort and device reset functions use them.
50 * */
51
52 /* CHANGELOG
53 *
54 * Version 2.8
55 *
56 * Fixed bad bug affecting tag starvation processing (previously the
57 * driver would hang the system if too many tags starved. Also fixed
58 * bad bug having to do with 10 byte command processing and REQUEST
59 * SENSE (the command would loop forever getting a transfer length
60 * mismatch in the CMD phase).
61 *
62 * Version 2.7
63 *
64 * Fixed scripts problem which caused certain devices (notably CDRWs)
65 * to hang on initial INQUIRY. Updated NCR_700_readl/writel to use
66 * __raw_readl/writel for parisc compatibility (Thomas
67 * Bogendoerfer). Added missing SCp->request_bufflen initialisation
68 * for sense requests (Ryan Bradetich).
69 *
70 * Version 2.6
71 *
72 * Following test of the 64 bit parisc kernel by Richard Hirst,
73 * several problems have now been corrected. Also adds support for
74 * consistent memory allocation.
75 *
76 * Version 2.5
77 *
78 * More Compatibility changes for 710 (now actually works). Enhanced
79 * support for odd clock speeds which constrain SDTR negotiations.
80 * correct cacheline separation for scsi messages and status for
81 * incoherent architectures. Use of the pci mapping functions on
82 * buffers to begin support for 64 bit drivers.
83 *
84 * Version 2.4
85 *
86 * Added support for the 53c710 chip (in 53c700 emulation mode only---no
87 * special 53c710 instructions or registers are used).
88 *
89 * Version 2.3
90 *
91 * More endianness/cache coherency changes.
92 *
93 * Better bad device handling (handles devices lying about tag
94 * queueing support and devices which fail to provide sense data on
95 * contingent allegiance conditions)
96 *
97 * Many thanks to Richard Hirst <rhirst@linuxcare.com> for patiently
98 * debugging this driver on the parisc architecture and suggesting
99 * many improvements and bug fixes.
100 *
101 * Thanks also go to Linuxcare Inc. for providing several PARISC
102 * machines for me to debug the driver on.
103 *
104 * Version 2.2
105 *
106 * Made the driver mem or io mapped; added endian invariance; added
107 * dma cache flushing operations for architectures which need it;
108 * added support for more varied clocking speeds.
109 *
110 * Version 2.1
111 *
112 * Initial modularisation from the D700. See NCR_D700.c for the rest of
113 * the changelog.
114 * */
115 #define NCR_700_VERSION "2.8"
116
117 #include <linux/kernel.h>
118 #include <linux/types.h>
119 #include <linux/string.h>
120 #include <linux/slab.h>
121 #include <linux/ioport.h>
122 #include <linux/delay.h>
123 #include <linux/spinlock.h>
124 #include <linux/completion.h>
125 #include <linux/init.h>
126 #include <linux/proc_fs.h>
127 #include <linux/blkdev.h>
128 #include <linux/module.h>
129 #include <linux/interrupt.h>
130 #include <linux/device.h>
131 #include <asm/dma.h>
132 #include <asm/io.h>
133 #include <asm/pgtable.h>
134 #include <asm/byteorder.h>
135
136 #include <scsi/scsi.h>
137 #include <scsi/scsi_cmnd.h>
138 #include <scsi/scsi_dbg.h>
139 #include <scsi/scsi_eh.h>
140 #include <scsi/scsi_host.h>
141 #include <scsi/scsi_tcq.h>
142 #include <scsi/scsi_transport.h>
143 #include <scsi/scsi_transport_spi.h>
144
145 #include "53c700.h"
146
147 /* NOTE: For 64 bit drivers there are points in the code where we use
148 * a non dereferenceable pointer to point to a structure in dma-able
149 * memory (which is 32 bits) so that we can use all of the structure
150 * operations but take the address at the end. This macro allows us
151 * to truncate the 64 bit pointer down to 32 bits without the compiler
152 * complaining */
153 #define to32bit(x) ((__u32)((unsigned long)(x)))
154
155 #ifdef NCR_700_DEBUG
156 #define STATIC
157 #else
158 #define STATIC static
159 #endif
160
161 MODULE_AUTHOR("James Bottomley");
162 MODULE_DESCRIPTION("53c700 and 53c700-66 Driver");
163 MODULE_LICENSE("GPL");
164
165 /* This is the script */
166 #include "53c700_d.h"
167
168
169 STATIC int NCR_700_queuecommand(struct Scsi_Host *h, struct scsi_cmnd *);
170 STATIC int NCR_700_abort(struct scsi_cmnd * SCpnt);
171 STATIC int NCR_700_bus_reset(struct scsi_cmnd * SCpnt);
172 STATIC int NCR_700_host_reset(struct scsi_cmnd * SCpnt);
173 STATIC void NCR_700_chip_setup(struct Scsi_Host *host);
174 STATIC void NCR_700_chip_reset(struct Scsi_Host *host);
175 STATIC int NCR_700_slave_alloc(struct scsi_device *SDpnt);
176 STATIC int NCR_700_slave_configure(struct scsi_device *SDpnt);
177 STATIC void NCR_700_slave_destroy(struct scsi_device *SDpnt);
178 static int NCR_700_change_queue_depth(struct scsi_device *SDpnt, int depth);
179 static int NCR_700_change_queue_type(struct scsi_device *SDpnt, int depth);
180
181 STATIC struct device_attribute *NCR_700_dev_attrs[];
182
183 STATIC struct scsi_transport_template *NCR_700_transport_template = NULL;
184
185 static char *NCR_700_phase[] = {
186 "",
187 "after selection",
188 "before command phase",
189 "after command phase",
190 "after status phase",
191 "after data in phase",
192 "after data out phase",
193 "during data phase",
194 };
195
196 static char *NCR_700_condition[] = {
197 "",
198 "NOT MSG_OUT",
199 "UNEXPECTED PHASE",
200 "NOT MSG_IN",
201 "UNEXPECTED MSG",
202 "MSG_IN",
203 "SDTR_MSG RECEIVED",
204 "REJECT_MSG RECEIVED",
205 "DISCONNECT_MSG RECEIVED",
206 "MSG_OUT",
207 "DATA_IN",
208
209 };
210
211 static char *NCR_700_fatal_messages[] = {
212 "unexpected message after reselection",
213 "still MSG_OUT after message injection",
214 "not MSG_IN after selection",
215 "Illegal message length received",
216 };
217
218 static char *NCR_700_SBCL_bits[] = {
219 "IO ",
220 "CD ",
221 "MSG ",
222 "ATN ",
223 "SEL ",
224 "BSY ",
225 "ACK ",
226 "REQ ",
227 };
228
229 static char *NCR_700_SBCL_to_phase[] = {
230 "DATA_OUT",
231 "DATA_IN",
232 "CMD_OUT",
233 "STATE",
234 "ILLEGAL PHASE",
235 "ILLEGAL PHASE",
236 "MSG OUT",
237 "MSG IN",
238 };
239
240 /* This translates the SDTR message offset and period to a value
241 * which can be loaded into the SXFER_REG.
242 *
243 * NOTE: According to SCSI-2, the true transfer period (in ns) is
244 * actually four times this period value */
245 static inline __u8
246 NCR_700_offset_period_to_sxfer(struct NCR_700_Host_Parameters *hostdata,
247 __u8 offset, __u8 period)
248 {
249 int XFERP;
250
251 __u8 min_xferp = (hostdata->chip710
252 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
253 __u8 max_offset = (hostdata->chip710
254 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET);
255
256 if(offset == 0)
257 return 0;
258
259 if(period < hostdata->min_period) {
260 printk(KERN_WARNING "53c700: Period %dns is less than this chip's minimum, setting to %d\n", period*4, NCR_700_MIN_PERIOD*4);
261 period = hostdata->min_period;
262 }
263 XFERP = (period*4 * hostdata->sync_clock)/1000 - 4;
264 if(offset > max_offset) {
265 printk(KERN_WARNING "53c700: Offset %d exceeds chip maximum, setting to %d\n",
266 offset, max_offset);
267 offset = max_offset;
268 }
269 if(XFERP < min_xferp) {
270 XFERP = min_xferp;
271 }
272 return (offset & 0x0f) | (XFERP & 0x07)<<4;
273 }
274
275 static inline __u8
276 NCR_700_get_SXFER(struct scsi_device *SDp)
277 {
278 struct NCR_700_Host_Parameters *hostdata =
279 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
280
281 return NCR_700_offset_period_to_sxfer(hostdata,
282 spi_offset(SDp->sdev_target),
283 spi_period(SDp->sdev_target));
284 }
285
286 struct Scsi_Host *
287 NCR_700_detect(struct scsi_host_template *tpnt,
288 struct NCR_700_Host_Parameters *hostdata, struct device *dev)
289 {
290 dma_addr_t pScript, pSlots;
291 __u8 *memory;
292 __u32 *script;
293 struct Scsi_Host *host;
294 static int banner = 0;
295 int j;
296
297 if(tpnt->sdev_attrs == NULL)
298 tpnt->sdev_attrs = NCR_700_dev_attrs;
299
300 memory = dma_alloc_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
301 &pScript, GFP_KERNEL);
302 if(memory == NULL) {
303 printk(KERN_ERR "53c700: Failed to allocate memory for driver, detatching\n");
304 return NULL;
305 }
306
307 script = (__u32 *)memory;
308 hostdata->msgin = memory + MSGIN_OFFSET;
309 hostdata->msgout = memory + MSGOUT_OFFSET;
310 hostdata->status = memory + STATUS_OFFSET;
311 hostdata->slots = (struct NCR_700_command_slot *)(memory + SLOTS_OFFSET);
312 hostdata->dev = dev;
313
314 pSlots = pScript + SLOTS_OFFSET;
315
316 /* Fill in the missing routines from the host template */
317 tpnt->queuecommand = NCR_700_queuecommand;
318 tpnt->eh_abort_handler = NCR_700_abort;
319 tpnt->eh_bus_reset_handler = NCR_700_bus_reset;
320 tpnt->eh_host_reset_handler = NCR_700_host_reset;
321 tpnt->can_queue = NCR_700_COMMAND_SLOTS_PER_HOST;
322 tpnt->sg_tablesize = NCR_700_SG_SEGMENTS;
323 tpnt->cmd_per_lun = NCR_700_CMD_PER_LUN;
324 tpnt->use_clustering = ENABLE_CLUSTERING;
325 tpnt->slave_configure = NCR_700_slave_configure;
326 tpnt->slave_destroy = NCR_700_slave_destroy;
327 tpnt->slave_alloc = NCR_700_slave_alloc;
328 tpnt->change_queue_depth = NCR_700_change_queue_depth;
329 tpnt->change_queue_type = NCR_700_change_queue_type;
330 tpnt->use_blk_tags = 1;
331
332 if(tpnt->name == NULL)
333 tpnt->name = "53c700";
334 if(tpnt->proc_name == NULL)
335 tpnt->proc_name = "53c700";
336
337 host = scsi_host_alloc(tpnt, 4);
338 if (!host)
339 return NULL;
340 memset(hostdata->slots, 0, sizeof(struct NCR_700_command_slot)
341 * NCR_700_COMMAND_SLOTS_PER_HOST);
342 for (j = 0; j < NCR_700_COMMAND_SLOTS_PER_HOST; j++) {
343 dma_addr_t offset = (dma_addr_t)((unsigned long)&hostdata->slots[j].SG[0]
344 - (unsigned long)&hostdata->slots[0].SG[0]);
345 hostdata->slots[j].pSG = (struct NCR_700_SG_List *)((unsigned long)(pSlots + offset));
346 if(j == 0)
347 hostdata->free_list = &hostdata->slots[j];
348 else
349 hostdata->slots[j-1].ITL_forw = &hostdata->slots[j];
350 hostdata->slots[j].state = NCR_700_SLOT_FREE;
351 }
352
353 for (j = 0; j < ARRAY_SIZE(SCRIPT); j++)
354 script[j] = bS_to_host(SCRIPT[j]);
355
356 /* adjust all labels to be bus physical */
357 for (j = 0; j < PATCHES; j++)
358 script[LABELPATCHES[j]] = bS_to_host(pScript + SCRIPT[LABELPATCHES[j]]);
359 /* now patch up fixed addresses. */
360 script_patch_32(hostdata->dev, script, MessageLocation,
361 pScript + MSGOUT_OFFSET);
362 script_patch_32(hostdata->dev, script, StatusAddress,
363 pScript + STATUS_OFFSET);
364 script_patch_32(hostdata->dev, script, ReceiveMsgAddress,
365 pScript + MSGIN_OFFSET);
366
367 hostdata->script = script;
368 hostdata->pScript = pScript;
369 dma_sync_single_for_device(hostdata->dev, pScript, sizeof(SCRIPT), DMA_TO_DEVICE);
370 hostdata->state = NCR_700_HOST_FREE;
371 hostdata->cmd = NULL;
372 host->max_id = 8;
373 host->max_lun = NCR_700_MAX_LUNS;
374 BUG_ON(NCR_700_transport_template == NULL);
375 host->transportt = NCR_700_transport_template;
376 host->unique_id = (unsigned long)hostdata->base;
377 hostdata->eh_complete = NULL;
378 host->hostdata[0] = (unsigned long)hostdata;
379 /* kick the chip */
380 NCR_700_writeb(0xff, host, CTEST9_REG);
381 if (hostdata->chip710)
382 hostdata->rev = (NCR_700_readb(host, CTEST8_REG)>>4) & 0x0f;
383 else
384 hostdata->rev = (NCR_700_readb(host, CTEST7_REG)>>4) & 0x0f;
385 hostdata->fast = (NCR_700_readb(host, CTEST9_REG) == 0);
386 if (banner == 0) {
387 printk(KERN_NOTICE "53c700: Version " NCR_700_VERSION " By James.Bottomley@HansenPartnership.com\n");
388 banner = 1;
389 }
390 printk(KERN_NOTICE "scsi%d: %s rev %d %s\n", host->host_no,
391 hostdata->chip710 ? "53c710" :
392 (hostdata->fast ? "53c700-66" : "53c700"),
393 hostdata->rev, hostdata->differential ?
394 "(Differential)" : "");
395 /* reset the chip */
396 NCR_700_chip_reset(host);
397
398 if (scsi_add_host(host, dev)) {
399 dev_printk(KERN_ERR, dev, "53c700: scsi_add_host failed\n");
400 scsi_host_put(host);
401 return NULL;
402 }
403
404 spi_signalling(host) = hostdata->differential ? SPI_SIGNAL_HVD :
405 SPI_SIGNAL_SE;
406
407 return host;
408 }
409
410 int
411 NCR_700_release(struct Scsi_Host *host)
412 {
413 struct NCR_700_Host_Parameters *hostdata =
414 (struct NCR_700_Host_Parameters *)host->hostdata[0];
415
416 dma_free_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
417 hostdata->script, hostdata->pScript);
418 return 1;
419 }
420
421 static inline __u8
422 NCR_700_identify(int can_disconnect, __u8 lun)
423 {
424 return IDENTIFY_BASE |
425 ((can_disconnect) ? 0x40 : 0) |
426 (lun & NCR_700_LUN_MASK);
427 }
428
429 /*
430 * Function : static int data_residual (Scsi_Host *host)
431 *
432 * Purpose : return residual data count of what's in the chip. If you
433 * really want to know what this function is doing, it's almost a
434 * direct transcription of the algorithm described in the 53c710
435 * guide, except that the DBC and DFIFO registers are only 6 bits
436 * wide on a 53c700.
437 *
438 * Inputs : host - SCSI host */
439 static inline int
440 NCR_700_data_residual (struct Scsi_Host *host) {
441 struct NCR_700_Host_Parameters *hostdata =
442 (struct NCR_700_Host_Parameters *)host->hostdata[0];
443 int count, synchronous = 0;
444 unsigned int ddir;
445
446 if(hostdata->chip710) {
447 count = ((NCR_700_readb(host, DFIFO_REG) & 0x7f) -
448 (NCR_700_readl(host, DBC_REG) & 0x7f)) & 0x7f;
449 } else {
450 count = ((NCR_700_readb(host, DFIFO_REG) & 0x3f) -
451 (NCR_700_readl(host, DBC_REG) & 0x3f)) & 0x3f;
452 }
453
454 if(hostdata->fast)
455 synchronous = NCR_700_readb(host, SXFER_REG) & 0x0f;
456
457 /* get the data direction */
458 ddir = NCR_700_readb(host, CTEST0_REG) & 0x01;
459
460 if (ddir) {
461 /* Receive */
462 if (synchronous)
463 count += (NCR_700_readb(host, SSTAT2_REG) & 0xf0) >> 4;
464 else
465 if (NCR_700_readb(host, SSTAT1_REG) & SIDL_REG_FULL)
466 ++count;
467 } else {
468 /* Send */
469 __u8 sstat = NCR_700_readb(host, SSTAT1_REG);
470 if (sstat & SODL_REG_FULL)
471 ++count;
472 if (synchronous && (sstat & SODR_REG_FULL))
473 ++count;
474 }
475 #ifdef NCR_700_DEBUG
476 if(count)
477 printk("RESIDUAL IS %d (ddir %d)\n", count, ddir);
478 #endif
479 return count;
480 }
481
482 /* print out the SCSI wires and corresponding phase from the SBCL register
483 * in the chip */
484 static inline char *
485 sbcl_to_string(__u8 sbcl)
486 {
487 int i;
488 static char ret[256];
489
490 ret[0]='\0';
491 for(i=0; i<8; i++) {
492 if((1<<i) & sbcl)
493 strcat(ret, NCR_700_SBCL_bits[i]);
494 }
495 strcat(ret, NCR_700_SBCL_to_phase[sbcl & 0x07]);
496 return ret;
497 }
498
499 static inline __u8
500 bitmap_to_number(__u8 bitmap)
501 {
502 __u8 i;
503
504 for(i=0; i<8 && !(bitmap &(1<<i)); i++)
505 ;
506 return i;
507 }
508
509 /* Pull a slot off the free list */
510 STATIC struct NCR_700_command_slot *
511 find_empty_slot(struct NCR_700_Host_Parameters *hostdata)
512 {
513 struct NCR_700_command_slot *slot = hostdata->free_list;
514
515 if(slot == NULL) {
516 /* sanity check */
517 if(hostdata->command_slot_count != NCR_700_COMMAND_SLOTS_PER_HOST)
518 printk(KERN_ERR "SLOTS FULL, but count is %d, should be %d\n", hostdata->command_slot_count, NCR_700_COMMAND_SLOTS_PER_HOST);
519 return NULL;
520 }
521
522 if(slot->state != NCR_700_SLOT_FREE)
523 /* should panic! */
524 printk(KERN_ERR "BUSY SLOT ON FREE LIST!!!\n");
525
526
527 hostdata->free_list = slot->ITL_forw;
528 slot->ITL_forw = NULL;
529
530
531 /* NOTE: set the state to busy here, not queued, since this
532 * indicates the slot is in use and cannot be run by the IRQ
533 * finish routine. If we cannot queue the command when it
534 * is properly build, we then change to NCR_700_SLOT_QUEUED */
535 slot->state = NCR_700_SLOT_BUSY;
536 slot->flags = 0;
537 hostdata->command_slot_count++;
538
539 return slot;
540 }
541
542 STATIC void
543 free_slot(struct NCR_700_command_slot *slot,
544 struct NCR_700_Host_Parameters *hostdata)
545 {
546 if((slot->state & NCR_700_SLOT_MASK) != NCR_700_SLOT_MAGIC) {
547 printk(KERN_ERR "53c700: SLOT %p is not MAGIC!!!\n", slot);
548 }
549 if(slot->state == NCR_700_SLOT_FREE) {
550 printk(KERN_ERR "53c700: SLOT %p is FREE!!!\n", slot);
551 }
552
553 slot->resume_offset = 0;
554 slot->cmnd = NULL;
555 slot->state = NCR_700_SLOT_FREE;
556 slot->ITL_forw = hostdata->free_list;
557 hostdata->free_list = slot;
558 hostdata->command_slot_count--;
559 }
560
561
562 /* This routine really does very little. The command is indexed on
563 the ITL and (if tagged) the ITLQ lists in _queuecommand */
564 STATIC void
565 save_for_reselection(struct NCR_700_Host_Parameters *hostdata,
566 struct scsi_cmnd *SCp, __u32 dsp)
567 {
568 /* Its just possible that this gets executed twice */
569 if(SCp != NULL) {
570 struct NCR_700_command_slot *slot =
571 (struct NCR_700_command_slot *)SCp->host_scribble;
572
573 slot->resume_offset = dsp;
574 }
575 hostdata->state = NCR_700_HOST_FREE;
576 hostdata->cmd = NULL;
577 }
578
579 STATIC inline void
580 NCR_700_unmap(struct NCR_700_Host_Parameters *hostdata, struct scsi_cmnd *SCp,
581 struct NCR_700_command_slot *slot)
582 {
583 if(SCp->sc_data_direction != DMA_NONE &&
584 SCp->sc_data_direction != DMA_BIDIRECTIONAL)
585 scsi_dma_unmap(SCp);
586 }
587
588 STATIC inline void
589 NCR_700_scsi_done(struct NCR_700_Host_Parameters *hostdata,
590 struct scsi_cmnd *SCp, int result)
591 {
592 hostdata->state = NCR_700_HOST_FREE;
593 hostdata->cmd = NULL;
594
595 if(SCp != NULL) {
596 struct NCR_700_command_slot *slot =
597 (struct NCR_700_command_slot *)SCp->host_scribble;
598
599 dma_unmap_single(hostdata->dev, slot->pCmd,
600 MAX_COMMAND_SIZE, DMA_TO_DEVICE);
601 if (slot->flags == NCR_700_FLAG_AUTOSENSE) {
602 char *cmnd = NCR_700_get_sense_cmnd(SCp->device);
603
604 dma_unmap_single(hostdata->dev, slot->dma_handle,
605 SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE);
606 /* restore the old result if the request sense was
607 * successful */
608 if (result == 0)
609 result = cmnd[7];
610 /* restore the original length */
611 SCp->cmd_len = cmnd[8];
612 } else
613 NCR_700_unmap(hostdata, SCp, slot);
614
615 free_slot(slot, hostdata);
616 #ifdef NCR_700_DEBUG
617 if(NCR_700_get_depth(SCp->device) == 0 ||
618 NCR_700_get_depth(SCp->device) > SCp->device->queue_depth)
619 printk(KERN_ERR "Invalid depth in NCR_700_scsi_done(): %d\n",
620 NCR_700_get_depth(SCp->device));
621 #endif /* NCR_700_DEBUG */
622 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) - 1);
623
624 SCp->host_scribble = NULL;
625 SCp->result = result;
626 SCp->scsi_done(SCp);
627 } else {
628 printk(KERN_ERR "53c700: SCSI DONE HAS NULL SCp\n");
629 }
630 }
631
632
633 STATIC void
634 NCR_700_internal_bus_reset(struct Scsi_Host *host)
635 {
636 /* Bus reset */
637 NCR_700_writeb(ASSERT_RST, host, SCNTL1_REG);
638 udelay(50);
639 NCR_700_writeb(0, host, SCNTL1_REG);
640
641 }
642
643 STATIC void
644 NCR_700_chip_setup(struct Scsi_Host *host)
645 {
646 struct NCR_700_Host_Parameters *hostdata =
647 (struct NCR_700_Host_Parameters *)host->hostdata[0];
648 __u8 min_period;
649 __u8 min_xferp = (hostdata->chip710 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
650
651 if(hostdata->chip710) {
652 __u8 burst_disable = 0;
653 __u8 burst_length = 0;
654
655 switch (hostdata->burst_length) {
656 case 1:
657 burst_length = BURST_LENGTH_1;
658 break;
659 case 2:
660 burst_length = BURST_LENGTH_2;
661 break;
662 case 4:
663 burst_length = BURST_LENGTH_4;
664 break;
665 case 8:
666 burst_length = BURST_LENGTH_8;
667 break;
668 default:
669 burst_disable = BURST_DISABLE;
670 break;
671 }
672 hostdata->dcntl_extra |= COMPAT_700_MODE;
673
674 NCR_700_writeb(hostdata->dcntl_extra, host, DCNTL_REG);
675 NCR_700_writeb(burst_length | hostdata->dmode_extra,
676 host, DMODE_710_REG);
677 NCR_700_writeb(burst_disable | hostdata->ctest7_extra |
678 (hostdata->differential ? DIFF : 0),
679 host, CTEST7_REG);
680 NCR_700_writeb(BTB_TIMER_DISABLE, host, CTEST0_REG);
681 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY | PARITY
682 | AUTO_ATN, host, SCNTL0_REG);
683 } else {
684 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra,
685 host, DMODE_700_REG);
686 NCR_700_writeb(hostdata->differential ?
687 DIFF : 0, host, CTEST7_REG);
688 if(hostdata->fast) {
689 /* this is for 700-66, does nothing on 700 */
690 NCR_700_writeb(LAST_DIS_ENBL | ENABLE_ACTIVE_NEGATION
691 | GENERATE_RECEIVE_PARITY, host,
692 CTEST8_REG);
693 } else {
694 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY
695 | PARITY | AUTO_ATN, host, SCNTL0_REG);
696 }
697 }
698
699 NCR_700_writeb(1 << host->this_id, host, SCID_REG);
700 NCR_700_writeb(0, host, SBCL_REG);
701 NCR_700_writeb(ASYNC_OPERATION, host, SXFER_REG);
702
703 NCR_700_writeb(PHASE_MM_INT | SEL_TIMEOUT_INT | GROSS_ERR_INT | UX_DISC_INT
704 | RST_INT | PAR_ERR_INT | SELECT_INT, host, SIEN_REG);
705
706 NCR_700_writeb(ABORT_INT | INT_INST_INT | ILGL_INST_INT, host, DIEN_REG);
707 NCR_700_writeb(ENABLE_SELECT, host, SCNTL1_REG);
708 if(hostdata->clock > 75) {
709 printk(KERN_ERR "53c700: Clock speed %dMHz is too high: 75Mhz is the maximum this chip can be driven at\n", hostdata->clock);
710 /* do the best we can, but the async clock will be out
711 * of spec: sync divider 2, async divider 3 */
712 DEBUG(("53c700: sync 2 async 3\n"));
713 NCR_700_writeb(SYNC_DIV_2_0, host, SBCL_REG);
714 NCR_700_writeb(ASYNC_DIV_3_0 | hostdata->dcntl_extra, host, DCNTL_REG);
715 hostdata->sync_clock = hostdata->clock/2;
716 } else if(hostdata->clock > 50 && hostdata->clock <= 75) {
717 /* sync divider 1.5, async divider 3 */
718 DEBUG(("53c700: sync 1.5 async 3\n"));
719 NCR_700_writeb(SYNC_DIV_1_5, host, SBCL_REG);
720 NCR_700_writeb(ASYNC_DIV_3_0 | hostdata->dcntl_extra, host, DCNTL_REG);
721 hostdata->sync_clock = hostdata->clock*2;
722 hostdata->sync_clock /= 3;
723
724 } else if(hostdata->clock > 37 && hostdata->clock <= 50) {
725 /* sync divider 1, async divider 2 */
726 DEBUG(("53c700: sync 1 async 2\n"));
727 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
728 NCR_700_writeb(ASYNC_DIV_2_0 | hostdata->dcntl_extra, host, DCNTL_REG);
729 hostdata->sync_clock = hostdata->clock;
730 } else if(hostdata->clock > 25 && hostdata->clock <=37) {
731 /* sync divider 1, async divider 1.5 */
732 DEBUG(("53c700: sync 1 async 1.5\n"));
733 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
734 NCR_700_writeb(ASYNC_DIV_1_5 | hostdata->dcntl_extra, host, DCNTL_REG);
735 hostdata->sync_clock = hostdata->clock;
736 } else {
737 DEBUG(("53c700: sync 1 async 1\n"));
738 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
739 NCR_700_writeb(ASYNC_DIV_1_0 | hostdata->dcntl_extra, host, DCNTL_REG);
740 /* sync divider 1, async divider 1 */
741 hostdata->sync_clock = hostdata->clock;
742 }
743 /* Calculate the actual minimum period that can be supported
744 * by our synchronous clock speed. See the 710 manual for
745 * exact details of this calculation which is based on a
746 * setting of the SXFER register */
747 min_period = 1000*(4+min_xferp)/(4*hostdata->sync_clock);
748 hostdata->min_period = NCR_700_MIN_PERIOD;
749 if(min_period > NCR_700_MIN_PERIOD)
750 hostdata->min_period = min_period;
751 }
752
753 STATIC void
754 NCR_700_chip_reset(struct Scsi_Host *host)
755 {
756 struct NCR_700_Host_Parameters *hostdata =
757 (struct NCR_700_Host_Parameters *)host->hostdata[0];
758 if(hostdata->chip710) {
759 NCR_700_writeb(SOFTWARE_RESET_710, host, ISTAT_REG);
760 udelay(100);
761
762 NCR_700_writeb(0, host, ISTAT_REG);
763 } else {
764 NCR_700_writeb(SOFTWARE_RESET, host, DCNTL_REG);
765 udelay(100);
766
767 NCR_700_writeb(0, host, DCNTL_REG);
768 }
769
770 mdelay(1000);
771
772 NCR_700_chip_setup(host);
773 }
774
775 /* The heart of the message processing engine is that the instruction
776 * immediately after the INT is the normal case (and so must be CLEAR
777 * ACK). If we want to do something else, we call that routine in
778 * scripts and set temp to be the normal case + 8 (skipping the CLEAR
779 * ACK) so that the routine returns correctly to resume its activity
780 * */
781 STATIC __u32
782 process_extended_message(struct Scsi_Host *host,
783 struct NCR_700_Host_Parameters *hostdata,
784 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
785 {
786 __u32 resume_offset = dsp, temp = dsp + 8;
787 __u8 pun = 0xff, lun = 0xff;
788
789 if(SCp != NULL) {
790 pun = SCp->device->id;
791 lun = SCp->device->lun;
792 }
793
794 switch(hostdata->msgin[2]) {
795 case A_SDTR_MSG:
796 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
797 struct scsi_target *starget = SCp->device->sdev_target;
798 __u8 period = hostdata->msgin[3];
799 __u8 offset = hostdata->msgin[4];
800
801 if(offset == 0 || period == 0) {
802 offset = 0;
803 period = 0;
804 }
805
806 spi_offset(starget) = offset;
807 spi_period(starget) = period;
808
809 if(NCR_700_is_flag_set(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION)) {
810 spi_display_xfer_agreement(starget);
811 NCR_700_clear_flag(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION);
812 }
813
814 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
815 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
816
817 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
818 host, SXFER_REG);
819
820 } else {
821 /* SDTR message out of the blue, reject it */
822 shost_printk(KERN_WARNING, host,
823 "Unexpected SDTR msg\n");
824 hostdata->msgout[0] = A_REJECT_MSG;
825 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE);
826 script_patch_16(hostdata->dev, hostdata->script,
827 MessageCount, 1);
828 /* SendMsgOut returns, so set up the return
829 * address */
830 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
831 }
832 break;
833
834 case A_WDTR_MSG:
835 printk(KERN_INFO "scsi%d: (%d:%d), Unsolicited WDTR after CMD, Rejecting\n",
836 host->host_no, pun, lun);
837 hostdata->msgout[0] = A_REJECT_MSG;
838 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE);
839 script_patch_16(hostdata->dev, hostdata->script, MessageCount,
840 1);
841 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
842
843 break;
844
845 default:
846 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
847 host->host_no, pun, lun,
848 NCR_700_phase[(dsps & 0xf00) >> 8]);
849 spi_print_msg(hostdata->msgin);
850 printk("\n");
851 /* just reject it */
852 hostdata->msgout[0] = A_REJECT_MSG;
853 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE);
854 script_patch_16(hostdata->dev, hostdata->script, MessageCount,
855 1);
856 /* SendMsgOut returns, so set up the return
857 * address */
858 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
859 }
860 NCR_700_writel(temp, host, TEMP_REG);
861 return resume_offset;
862 }
863
864 STATIC __u32
865 process_message(struct Scsi_Host *host, struct NCR_700_Host_Parameters *hostdata,
866 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
867 {
868 /* work out where to return to */
869 __u32 temp = dsp + 8, resume_offset = dsp;
870 __u8 pun = 0xff, lun = 0xff;
871
872 if(SCp != NULL) {
873 pun = SCp->device->id;
874 lun = SCp->device->lun;
875 }
876
877 #ifdef NCR_700_DEBUG
878 printk("scsi%d (%d:%d): message %s: ", host->host_no, pun, lun,
879 NCR_700_phase[(dsps & 0xf00) >> 8]);
880 spi_print_msg(hostdata->msgin);
881 printk("\n");
882 #endif
883
884 switch(hostdata->msgin[0]) {
885
886 case A_EXTENDED_MSG:
887 resume_offset = process_extended_message(host, hostdata, SCp,
888 dsp, dsps);
889 break;
890
891 case A_REJECT_MSG:
892 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
893 /* Rejected our sync negotiation attempt */
894 spi_period(SCp->device->sdev_target) =
895 spi_offset(SCp->device->sdev_target) = 0;
896 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
897 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
898 } else if(SCp != NULL && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION) {
899 /* rejected our first simple tag message */
900 scmd_printk(KERN_WARNING, SCp,
901 "Rejected first tag queue attempt, turning off tag queueing\n");
902 /* we're done negotiating */
903 NCR_700_set_tag_neg_state(SCp->device, NCR_700_FINISHED_TAG_NEGOTIATION);
904 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
905
906 SCp->device->tagged_supported = 0;
907 scsi_change_queue_depth(SCp->device, host->cmd_per_lun);
908 scsi_set_tag_type(SCp->device, 0);
909 } else {
910 shost_printk(KERN_WARNING, host,
911 "(%d:%d) Unexpected REJECT Message %s\n",
912 pun, lun,
913 NCR_700_phase[(dsps & 0xf00) >> 8]);
914 /* however, just ignore it */
915 }
916 break;
917
918 case A_PARITY_ERROR_MSG:
919 printk(KERN_ERR "scsi%d (%d:%d) Parity Error!\n", host->host_no,
920 pun, lun);
921 NCR_700_internal_bus_reset(host);
922 break;
923 case A_SIMPLE_TAG_MSG:
924 printk(KERN_INFO "scsi%d (%d:%d) SIMPLE TAG %d %s\n", host->host_no,
925 pun, lun, hostdata->msgin[1],
926 NCR_700_phase[(dsps & 0xf00) >> 8]);
927 /* just ignore it */
928 break;
929 default:
930 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
931 host->host_no, pun, lun,
932 NCR_700_phase[(dsps & 0xf00) >> 8]);
933
934 spi_print_msg(hostdata->msgin);
935 printk("\n");
936 /* just reject it */
937 hostdata->msgout[0] = A_REJECT_MSG;
938 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE);
939 script_patch_16(hostdata->dev, hostdata->script, MessageCount,
940 1);
941 /* SendMsgOut returns, so set up the return
942 * address */
943 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
944
945 break;
946 }
947 NCR_700_writel(temp, host, TEMP_REG);
948 /* set us up to receive another message */
949 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
950 return resume_offset;
951 }
952
953 STATIC __u32
954 process_script_interrupt(__u32 dsps, __u32 dsp, struct scsi_cmnd *SCp,
955 struct Scsi_Host *host,
956 struct NCR_700_Host_Parameters *hostdata)
957 {
958 __u32 resume_offset = 0;
959 __u8 pun = 0xff, lun=0xff;
960
961 if(SCp != NULL) {
962 pun = SCp->device->id;
963 lun = SCp->device->lun;
964 }
965
966 if(dsps == A_GOOD_STATUS_AFTER_STATUS) {
967 DEBUG((" COMMAND COMPLETE, status=%02x\n",
968 hostdata->status[0]));
969 /* OK, if TCQ still under negotiation, we now know it works */
970 if (NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION)
971 NCR_700_set_tag_neg_state(SCp->device,
972 NCR_700_FINISHED_TAG_NEGOTIATION);
973
974 /* check for contingent allegiance contitions */
975 if(status_byte(hostdata->status[0]) == CHECK_CONDITION ||
976 status_byte(hostdata->status[0]) == COMMAND_TERMINATED) {
977 struct NCR_700_command_slot *slot =
978 (struct NCR_700_command_slot *)SCp->host_scribble;
979 if(slot->flags == NCR_700_FLAG_AUTOSENSE) {
980 /* OOPS: bad device, returning another
981 * contingent allegiance condition */
982 scmd_printk(KERN_ERR, SCp,
983 "broken device is looping in contingent allegiance: ignoring\n");
984 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
985 } else {
986 char *cmnd =
987 NCR_700_get_sense_cmnd(SCp->device);
988 #ifdef NCR_DEBUG
989 scsi_print_command(SCp);
990 printk(" cmd %p has status %d, requesting sense\n",
991 SCp, hostdata->status[0]);
992 #endif
993 /* we can destroy the command here
994 * because the contingent allegiance
995 * condition will cause a retry which
996 * will re-copy the command from the
997 * saved data_cmnd. We also unmap any
998 * data associated with the command
999 * here */
1000 NCR_700_unmap(hostdata, SCp, slot);
1001 dma_unmap_single(hostdata->dev, slot->pCmd,
1002 MAX_COMMAND_SIZE,
1003 DMA_TO_DEVICE);
1004
1005 cmnd[0] = REQUEST_SENSE;
1006 cmnd[1] = (lun & 0x7) << 5;
1007 cmnd[2] = 0;
1008 cmnd[3] = 0;
1009 cmnd[4] = SCSI_SENSE_BUFFERSIZE;
1010 cmnd[5] = 0;
1011 /* Here's a quiet hack: the
1012 * REQUEST_SENSE command is six bytes,
1013 * so store a flag indicating that
1014 * this was an internal sense request
1015 * and the original status at the end
1016 * of the command */
1017 cmnd[6] = NCR_700_INTERNAL_SENSE_MAGIC;
1018 cmnd[7] = hostdata->status[0];
1019 cmnd[8] = SCp->cmd_len;
1020 SCp->cmd_len = 6; /* command length for
1021 * REQUEST_SENSE */
1022 slot->pCmd = dma_map_single(hostdata->dev, cmnd, MAX_COMMAND_SIZE, DMA_TO_DEVICE);
1023 slot->dma_handle = dma_map_single(hostdata->dev, SCp->sense_buffer, SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE);
1024 slot->SG[0].ins = bS_to_host(SCRIPT_MOVE_DATA_IN | SCSI_SENSE_BUFFERSIZE);
1025 slot->SG[0].pAddr = bS_to_host(slot->dma_handle);
1026 slot->SG[1].ins = bS_to_host(SCRIPT_RETURN);
1027 slot->SG[1].pAddr = 0;
1028 slot->resume_offset = hostdata->pScript;
1029 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG[0])*2, DMA_TO_DEVICE);
1030 dma_cache_sync(hostdata->dev, SCp->sense_buffer, SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE);
1031
1032 /* queue the command for reissue */
1033 slot->state = NCR_700_SLOT_QUEUED;
1034 slot->flags = NCR_700_FLAG_AUTOSENSE;
1035 hostdata->state = NCR_700_HOST_FREE;
1036 hostdata->cmd = NULL;
1037 }
1038 } else {
1039 // Currently rely on the mid layer evaluation
1040 // of the tag queuing capability
1041 //
1042 //if(status_byte(hostdata->status[0]) == GOOD &&
1043 // SCp->cmnd[0] == INQUIRY && SCp->use_sg == 0) {
1044 // /* Piggy back the tag queueing support
1045 // * on this command */
1046 // dma_sync_single_for_cpu(hostdata->dev,
1047 // slot->dma_handle,
1048 // SCp->request_bufflen,
1049 // DMA_FROM_DEVICE);
1050 // if(((char *)SCp->request_buffer)[7] & 0x02) {
1051 // scmd_printk(KERN_INFO, SCp,
1052 // "Enabling Tag Command Queuing\n");
1053 // hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1054 // NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
1055 // } else {
1056 // NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
1057 // hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1058 // }
1059 //}
1060 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
1061 }
1062 } else if((dsps & 0xfffff0f0) == A_UNEXPECTED_PHASE) {
1063 __u8 i = (dsps & 0xf00) >> 8;
1064
1065 scmd_printk(KERN_ERR, SCp, "UNEXPECTED PHASE %s (%s)\n",
1066 NCR_700_phase[i],
1067 sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
1068 scmd_printk(KERN_ERR, SCp, " len = %d, cmd =",
1069 SCp->cmd_len);
1070 scsi_print_command(SCp);
1071
1072 NCR_700_internal_bus_reset(host);
1073 } else if((dsps & 0xfffff000) == A_FATAL) {
1074 int i = (dsps & 0xfff);
1075
1076 printk(KERN_ERR "scsi%d: (%d:%d) FATAL ERROR: %s\n",
1077 host->host_no, pun, lun, NCR_700_fatal_messages[i]);
1078 if(dsps == A_FATAL_ILLEGAL_MSG_LENGTH) {
1079 printk(KERN_ERR " msg begins %02x %02x\n",
1080 hostdata->msgin[0], hostdata->msgin[1]);
1081 }
1082 NCR_700_internal_bus_reset(host);
1083 } else if((dsps & 0xfffff0f0) == A_DISCONNECT) {
1084 #ifdef NCR_700_DEBUG
1085 __u8 i = (dsps & 0xf00) >> 8;
1086
1087 printk("scsi%d: (%d:%d), DISCONNECTED (%d) %s\n",
1088 host->host_no, pun, lun,
1089 i, NCR_700_phase[i]);
1090 #endif
1091 save_for_reselection(hostdata, SCp, dsp);
1092
1093 } else if(dsps == A_RESELECTION_IDENTIFIED) {
1094 __u8 lun;
1095 struct NCR_700_command_slot *slot;
1096 __u8 reselection_id = hostdata->reselection_id;
1097 struct scsi_device *SDp;
1098
1099 lun = hostdata->msgin[0] & 0x1f;
1100
1101 hostdata->reselection_id = 0xff;
1102 DEBUG(("scsi%d: (%d:%d) RESELECTED!\n",
1103 host->host_no, reselection_id, lun));
1104 /* clear the reselection indicator */
1105 SDp = __scsi_device_lookup(host, 0, reselection_id, lun);
1106 if(unlikely(SDp == NULL)) {
1107 printk(KERN_ERR "scsi%d: (%d:%d) HAS NO device\n",
1108 host->host_no, reselection_id, lun);
1109 BUG();
1110 }
1111 if(hostdata->msgin[1] == A_SIMPLE_TAG_MSG) {
1112 struct scsi_cmnd *SCp = scsi_find_tag(SDp, hostdata->msgin[2]);
1113 if(unlikely(SCp == NULL)) {
1114 printk(KERN_ERR "scsi%d: (%d:%d) no saved request for tag %d\n",
1115 host->host_no, reselection_id, lun, hostdata->msgin[2]);
1116 BUG();
1117 }
1118
1119 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1120 DDEBUG(KERN_DEBUG, SDp,
1121 "reselection is tag %d, slot %p(%d)\n",
1122 hostdata->msgin[2], slot, slot->tag);
1123 } else {
1124 struct scsi_cmnd *SCp = scsi_find_tag(SDp, SCSI_NO_TAG);
1125 if(unlikely(SCp == NULL)) {
1126 sdev_printk(KERN_ERR, SDp,
1127 "no saved request for untagged cmd\n");
1128 BUG();
1129 }
1130 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1131 }
1132
1133 if(slot == NULL) {
1134 printk(KERN_ERR "scsi%d: (%d:%d) RESELECTED but no saved command (MSG = %02x %02x %02x)!!\n",
1135 host->host_no, reselection_id, lun,
1136 hostdata->msgin[0], hostdata->msgin[1],
1137 hostdata->msgin[2]);
1138 } else {
1139 if(hostdata->state != NCR_700_HOST_BUSY)
1140 printk(KERN_ERR "scsi%d: FATAL, host not busy during valid reselection!\n",
1141 host->host_no);
1142 resume_offset = slot->resume_offset;
1143 hostdata->cmd = slot->cmnd;
1144
1145 /* re-patch for this command */
1146 script_patch_32_abs(hostdata->dev, hostdata->script,
1147 CommandAddress, slot->pCmd);
1148 script_patch_16(hostdata->dev, hostdata->script,
1149 CommandCount, slot->cmnd->cmd_len);
1150 script_patch_32_abs(hostdata->dev, hostdata->script,
1151 SGScriptStartAddress,
1152 to32bit(&slot->pSG[0].ins));
1153
1154 /* Note: setting SXFER only works if we're
1155 * still in the MESSAGE phase, so it is vital
1156 * that ACK is still asserted when we process
1157 * the reselection message. The resume offset
1158 * should therefore always clear ACK */
1159 NCR_700_writeb(NCR_700_get_SXFER(hostdata->cmd->device),
1160 host, SXFER_REG);
1161 dma_cache_sync(hostdata->dev, hostdata->msgin,
1162 MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
1163 dma_cache_sync(hostdata->dev, hostdata->msgout,
1164 MSG_ARRAY_SIZE, DMA_TO_DEVICE);
1165 /* I'm just being paranoid here, the command should
1166 * already have been flushed from the cache */
1167 dma_cache_sync(hostdata->dev, slot->cmnd->cmnd,
1168 slot->cmnd->cmd_len, DMA_TO_DEVICE);
1169
1170
1171
1172 }
1173 } else if(dsps == A_RESELECTED_DURING_SELECTION) {
1174
1175 /* This section is full of debugging code because I've
1176 * never managed to reach it. I think what happens is
1177 * that, because the 700 runs with selection
1178 * interrupts enabled the whole time that we take a
1179 * selection interrupt before we manage to get to the
1180 * reselected script interrupt */
1181
1182 __u8 reselection_id = NCR_700_readb(host, SFBR_REG);
1183 struct NCR_700_command_slot *slot;
1184
1185 /* Take out our own ID */
1186 reselection_id &= ~(1<<host->this_id);
1187
1188 /* I've never seen this happen, so keep this as a printk rather
1189 * than a debug */
1190 printk(KERN_INFO "scsi%d: (%d:%d) RESELECTION DURING SELECTION, dsp=%08x[%04x] state=%d, count=%d\n",
1191 host->host_no, reselection_id, lun, dsp, dsp - hostdata->pScript, hostdata->state, hostdata->command_slot_count);
1192
1193 {
1194 /* FIXME: DEBUGGING CODE */
1195 __u32 SG = (__u32)bS_to_cpu(hostdata->script[A_SGScriptStartAddress_used[0]]);
1196 int i;
1197
1198 for(i=0; i< NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1199 if(SG >= to32bit(&hostdata->slots[i].pSG[0])
1200 && SG <= to32bit(&hostdata->slots[i].pSG[NCR_700_SG_SEGMENTS]))
1201 break;
1202 }
1203 printk(KERN_INFO "IDENTIFIED SG segment as being %08x in slot %p, cmd %p, slot->resume_offset=%08x\n", SG, &hostdata->slots[i], hostdata->slots[i].cmnd, hostdata->slots[i].resume_offset);
1204 SCp = hostdata->slots[i].cmnd;
1205 }
1206
1207 if(SCp != NULL) {
1208 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1209 /* change slot from busy to queued to redo command */
1210 slot->state = NCR_700_SLOT_QUEUED;
1211 }
1212 hostdata->cmd = NULL;
1213
1214 if(reselection_id == 0) {
1215 if(hostdata->reselection_id == 0xff) {
1216 printk(KERN_ERR "scsi%d: Invalid reselection during selection!!\n", host->host_no);
1217 return 0;
1218 } else {
1219 printk(KERN_ERR "scsi%d: script reselected and we took a selection interrupt\n",
1220 host->host_no);
1221 reselection_id = hostdata->reselection_id;
1222 }
1223 } else {
1224
1225 /* convert to real ID */
1226 reselection_id = bitmap_to_number(reselection_id);
1227 }
1228 hostdata->reselection_id = reselection_id;
1229 /* just in case we have a stale simple tag message, clear it */
1230 hostdata->msgin[1] = 0;
1231 dma_cache_sync(hostdata->dev, hostdata->msgin,
1232 MSG_ARRAY_SIZE, DMA_BIDIRECTIONAL);
1233 if(hostdata->tag_negotiated & (1<<reselection_id)) {
1234 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1235 } else {
1236 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1237 }
1238 } else if(dsps == A_COMPLETED_SELECTION_AS_TARGET) {
1239 /* we've just disconnected from the bus, do nothing since
1240 * a return here will re-run the queued command slot
1241 * that may have been interrupted by the initial selection */
1242 DEBUG((" SELECTION COMPLETED\n"));
1243 } else if((dsps & 0xfffff0f0) == A_MSG_IN) {
1244 resume_offset = process_message(host, hostdata, SCp,
1245 dsp, dsps);
1246 } else if((dsps & 0xfffff000) == 0) {
1247 __u8 i = (dsps & 0xf0) >> 4, j = (dsps & 0xf00) >> 8;
1248 printk(KERN_ERR "scsi%d: (%d:%d), unhandled script condition %s %s at %04x\n",
1249 host->host_no, pun, lun, NCR_700_condition[i],
1250 NCR_700_phase[j], dsp - hostdata->pScript);
1251 if(SCp != NULL) {
1252 struct scatterlist *sg;
1253
1254 scsi_print_command(SCp);
1255 scsi_for_each_sg(SCp, sg, scsi_sg_count(SCp) + 1, i) {
1256 printk(KERN_INFO " SG[%d].length = %d, move_insn=%08x, addr %08x\n", i, sg->length, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].ins, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].pAddr);
1257 }
1258 }
1259 NCR_700_internal_bus_reset(host);
1260 } else if((dsps & 0xfffff000) == A_DEBUG_INTERRUPT) {
1261 printk(KERN_NOTICE "scsi%d (%d:%d) DEBUG INTERRUPT %d AT %08x[%04x], continuing\n",
1262 host->host_no, pun, lun, dsps & 0xfff, dsp, dsp - hostdata->pScript);
1263 resume_offset = dsp;
1264 } else {
1265 printk(KERN_ERR "scsi%d: (%d:%d), unidentified script interrupt 0x%x at %04x\n",
1266 host->host_no, pun, lun, dsps, dsp - hostdata->pScript);
1267 NCR_700_internal_bus_reset(host);
1268 }
1269 return resume_offset;
1270 }
1271
1272 /* We run the 53c700 with selection interrupts always enabled. This
1273 * means that the chip may be selected as soon as the bus frees. On a
1274 * busy bus, this can be before the scripts engine finishes its
1275 * processing. Therefore, part of the selection processing has to be
1276 * to find out what the scripts engine is doing and complete the
1277 * function if necessary (i.e. process the pending disconnect or save
1278 * the interrupted initial selection */
1279 STATIC inline __u32
1280 process_selection(struct Scsi_Host *host, __u32 dsp)
1281 {
1282 __u8 id = 0; /* Squash compiler warning */
1283 int count = 0;
1284 __u32 resume_offset = 0;
1285 struct NCR_700_Host_Parameters *hostdata =
1286 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1287 struct scsi_cmnd *SCp = hostdata->cmd;
1288 __u8 sbcl;
1289
1290 for(count = 0; count < 5; count++) {
1291 id = NCR_700_readb(host, hostdata->chip710 ?
1292 CTEST9_REG : SFBR_REG);
1293
1294 /* Take out our own ID */
1295 id &= ~(1<<host->this_id);
1296 if(id != 0)
1297 break;
1298 udelay(5);
1299 }
1300 sbcl = NCR_700_readb(host, SBCL_REG);
1301 if((sbcl & SBCL_IO) == 0) {
1302 /* mark as having been selected rather than reselected */
1303 id = 0xff;
1304 } else {
1305 /* convert to real ID */
1306 hostdata->reselection_id = id = bitmap_to_number(id);
1307 DEBUG(("scsi%d: Reselected by %d\n",
1308 host->host_no, id));
1309 }
1310 if(hostdata->state == NCR_700_HOST_BUSY && SCp != NULL) {
1311 struct NCR_700_command_slot *slot =
1312 (struct NCR_700_command_slot *)SCp->host_scribble;
1313 DEBUG((" ID %d WARNING: RESELECTION OF BUSY HOST, saving cmd %p, slot %p, addr %x [%04x], resume %x!\n", id, hostdata->cmd, slot, dsp, dsp - hostdata->pScript, resume_offset));
1314
1315 switch(dsp - hostdata->pScript) {
1316 case Ent_Disconnect1:
1317 case Ent_Disconnect2:
1318 save_for_reselection(hostdata, SCp, Ent_Disconnect2 + hostdata->pScript);
1319 break;
1320 case Ent_Disconnect3:
1321 case Ent_Disconnect4:
1322 save_for_reselection(hostdata, SCp, Ent_Disconnect4 + hostdata->pScript);
1323 break;
1324 case Ent_Disconnect5:
1325 case Ent_Disconnect6:
1326 save_for_reselection(hostdata, SCp, Ent_Disconnect6 + hostdata->pScript);
1327 break;
1328 case Ent_Disconnect7:
1329 case Ent_Disconnect8:
1330 save_for_reselection(hostdata, SCp, Ent_Disconnect8 + hostdata->pScript);
1331 break;
1332 case Ent_Finish1:
1333 case Ent_Finish2:
1334 process_script_interrupt(A_GOOD_STATUS_AFTER_STATUS, dsp, SCp, host, hostdata);
1335 break;
1336
1337 default:
1338 slot->state = NCR_700_SLOT_QUEUED;
1339 break;
1340 }
1341 }
1342 hostdata->state = NCR_700_HOST_BUSY;
1343 hostdata->cmd = NULL;
1344 /* clear any stale simple tag message */
1345 hostdata->msgin[1] = 0;
1346 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE,
1347 DMA_BIDIRECTIONAL);
1348
1349 if(id == 0xff) {
1350 /* Selected as target, Ignore */
1351 resume_offset = hostdata->pScript + Ent_SelectedAsTarget;
1352 } else if(hostdata->tag_negotiated & (1<<id)) {
1353 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1354 } else {
1355 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1356 }
1357 return resume_offset;
1358 }
1359
1360 static inline void
1361 NCR_700_clear_fifo(struct Scsi_Host *host) {
1362 const struct NCR_700_Host_Parameters *hostdata
1363 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1364 if(hostdata->chip710) {
1365 NCR_700_writeb(CLR_FIFO_710, host, CTEST8_REG);
1366 } else {
1367 NCR_700_writeb(CLR_FIFO, host, DFIFO_REG);
1368 }
1369 }
1370
1371 static inline void
1372 NCR_700_flush_fifo(struct Scsi_Host *host) {
1373 const struct NCR_700_Host_Parameters *hostdata
1374 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1375 if(hostdata->chip710) {
1376 NCR_700_writeb(FLUSH_DMA_FIFO_710, host, CTEST8_REG);
1377 udelay(10);
1378 NCR_700_writeb(0, host, CTEST8_REG);
1379 } else {
1380 NCR_700_writeb(FLUSH_DMA_FIFO, host, DFIFO_REG);
1381 udelay(10);
1382 NCR_700_writeb(0, host, DFIFO_REG);
1383 }
1384 }
1385
1386
1387 /* The queue lock with interrupts disabled must be held on entry to
1388 * this function */
1389 STATIC int
1390 NCR_700_start_command(struct scsi_cmnd *SCp)
1391 {
1392 struct NCR_700_command_slot *slot =
1393 (struct NCR_700_command_slot *)SCp->host_scribble;
1394 struct NCR_700_Host_Parameters *hostdata =
1395 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1396 __u16 count = 1; /* for IDENTIFY message */
1397 u8 lun = SCp->device->lun;
1398
1399 if(hostdata->state != NCR_700_HOST_FREE) {
1400 /* keep this inside the lock to close the race window where
1401 * the running command finishes on another CPU while we don't
1402 * change the state to queued on this one */
1403 slot->state = NCR_700_SLOT_QUEUED;
1404
1405 DEBUG(("scsi%d: host busy, queueing command %p, slot %p\n",
1406 SCp->device->host->host_no, slot->cmnd, slot));
1407 return 0;
1408 }
1409 hostdata->state = NCR_700_HOST_BUSY;
1410 hostdata->cmd = SCp;
1411 slot->state = NCR_700_SLOT_BUSY;
1412 /* keep interrupts disabled until we have the command correctly
1413 * set up so we cannot take a selection interrupt */
1414
1415 hostdata->msgout[0] = NCR_700_identify((SCp->cmnd[0] != REQUEST_SENSE &&
1416 slot->flags != NCR_700_FLAG_AUTOSENSE),
1417 lun);
1418 /* for INQUIRY or REQUEST_SENSE commands, we cannot be sure
1419 * if the negotiated transfer parameters still hold, so
1420 * always renegotiate them */
1421 if(SCp->cmnd[0] == INQUIRY || SCp->cmnd[0] == REQUEST_SENSE ||
1422 slot->flags == NCR_700_FLAG_AUTOSENSE) {
1423 NCR_700_clear_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
1424 }
1425
1426 /* REQUEST_SENSE is asking for contingent I_T_L(_Q) status.
1427 * If a contingent allegiance condition exists, the device
1428 * will refuse all tags, so send the request sense as untagged
1429 * */
1430 if((hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1431 && (slot->tag != SCSI_NO_TAG && SCp->cmnd[0] != REQUEST_SENSE &&
1432 slot->flags != NCR_700_FLAG_AUTOSENSE)) {
1433 count += spi_populate_tag_msg(&hostdata->msgout[count], SCp);
1434 }
1435
1436 if(hostdata->fast &&
1437 NCR_700_is_flag_clear(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC)) {
1438 count += spi_populate_sync_msg(&hostdata->msgout[count],
1439 spi_period(SCp->device->sdev_target),
1440 spi_offset(SCp->device->sdev_target));
1441 NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
1442 }
1443
1444 script_patch_16(hostdata->dev, hostdata->script, MessageCount, count);
1445
1446
1447 script_patch_ID(hostdata->dev, hostdata->script,
1448 Device_ID, 1<<scmd_id(SCp));
1449
1450 script_patch_32_abs(hostdata->dev, hostdata->script, CommandAddress,
1451 slot->pCmd);
1452 script_patch_16(hostdata->dev, hostdata->script, CommandCount,
1453 SCp->cmd_len);
1454 /* finally plumb the beginning of the SG list into the script
1455 * */
1456 script_patch_32_abs(hostdata->dev, hostdata->script,
1457 SGScriptStartAddress, to32bit(&slot->pSG[0].ins));
1458 NCR_700_clear_fifo(SCp->device->host);
1459
1460 if(slot->resume_offset == 0)
1461 slot->resume_offset = hostdata->pScript;
1462 /* now perform all the writebacks and invalidates */
1463 dma_cache_sync(hostdata->dev, hostdata->msgout, count, DMA_TO_DEVICE);
1464 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE,
1465 DMA_FROM_DEVICE);
1466 dma_cache_sync(hostdata->dev, SCp->cmnd, SCp->cmd_len, DMA_TO_DEVICE);
1467 dma_cache_sync(hostdata->dev, hostdata->status, 1, DMA_FROM_DEVICE);
1468
1469 /* set the synchronous period/offset */
1470 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
1471 SCp->device->host, SXFER_REG);
1472 NCR_700_writel(slot->temp, SCp->device->host, TEMP_REG);
1473 NCR_700_writel(slot->resume_offset, SCp->device->host, DSP_REG);
1474
1475 return 1;
1476 }
1477
1478 irqreturn_t
1479 NCR_700_intr(int irq, void *dev_id)
1480 {
1481 struct Scsi_Host *host = (struct Scsi_Host *)dev_id;
1482 struct NCR_700_Host_Parameters *hostdata =
1483 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1484 __u8 istat;
1485 __u32 resume_offset = 0;
1486 __u8 pun = 0xff, lun = 0xff;
1487 unsigned long flags;
1488 int handled = 0;
1489
1490 /* Use the host lock to serialise access to the 53c700
1491 * hardware. Note: In future, we may need to take the queue
1492 * lock to enter the done routines. When that happens, we
1493 * need to ensure that for this driver, the host lock and the
1494 * queue lock point to the same thing. */
1495 spin_lock_irqsave(host->host_lock, flags);
1496 if((istat = NCR_700_readb(host, ISTAT_REG))
1497 & (SCSI_INT_PENDING | DMA_INT_PENDING)) {
1498 __u32 dsps;
1499 __u8 sstat0 = 0, dstat = 0;
1500 __u32 dsp;
1501 struct scsi_cmnd *SCp = hostdata->cmd;
1502 enum NCR_700_Host_State state;
1503
1504 handled = 1;
1505 state = hostdata->state;
1506 SCp = hostdata->cmd;
1507
1508 if(istat & SCSI_INT_PENDING) {
1509 udelay(10);
1510
1511 sstat0 = NCR_700_readb(host, SSTAT0_REG);
1512 }
1513
1514 if(istat & DMA_INT_PENDING) {
1515 udelay(10);
1516
1517 dstat = NCR_700_readb(host, DSTAT_REG);
1518 }
1519
1520 dsps = NCR_700_readl(host, DSPS_REG);
1521 dsp = NCR_700_readl(host, DSP_REG);
1522
1523 DEBUG(("scsi%d: istat %02x sstat0 %02x dstat %02x dsp %04x[%08x] dsps 0x%x\n",
1524 host->host_no, istat, sstat0, dstat,
1525 (dsp - (__u32)(hostdata->pScript))/4,
1526 dsp, dsps));
1527
1528 if(SCp != NULL) {
1529 pun = SCp->device->id;
1530 lun = SCp->device->lun;
1531 }
1532
1533 if(sstat0 & SCSI_RESET_DETECTED) {
1534 struct scsi_device *SDp;
1535 int i;
1536
1537 hostdata->state = NCR_700_HOST_BUSY;
1538
1539 printk(KERN_ERR "scsi%d: Bus Reset detected, executing command %p, slot %p, dsp %08x[%04x]\n",
1540 host->host_no, SCp, SCp == NULL ? NULL : SCp->host_scribble, dsp, dsp - hostdata->pScript);
1541
1542 scsi_report_bus_reset(host, 0);
1543
1544 /* clear all the negotiated parameters */
1545 __shost_for_each_device(SDp, host)
1546 NCR_700_clear_flag(SDp, ~0);
1547
1548 /* clear all the slots and their pending commands */
1549 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1550 struct scsi_cmnd *SCp;
1551 struct NCR_700_command_slot *slot =
1552 &hostdata->slots[i];
1553
1554 if(slot->state == NCR_700_SLOT_FREE)
1555 continue;
1556
1557 SCp = slot->cmnd;
1558 printk(KERN_ERR " failing command because of reset, slot %p, cmnd %p\n",
1559 slot, SCp);
1560 free_slot(slot, hostdata);
1561 SCp->host_scribble = NULL;
1562 NCR_700_set_depth(SCp->device, 0);
1563 /* NOTE: deadlock potential here: we
1564 * rely on mid-layer guarantees that
1565 * scsi_done won't try to issue the
1566 * command again otherwise we'll
1567 * deadlock on the
1568 * hostdata->state_lock */
1569 SCp->result = DID_RESET << 16;
1570 SCp->scsi_done(SCp);
1571 }
1572 mdelay(25);
1573 NCR_700_chip_setup(host);
1574
1575 hostdata->state = NCR_700_HOST_FREE;
1576 hostdata->cmd = NULL;
1577 /* signal back if this was an eh induced reset */
1578 if(hostdata->eh_complete != NULL)
1579 complete(hostdata->eh_complete);
1580 goto out_unlock;
1581 } else if(sstat0 & SELECTION_TIMEOUT) {
1582 DEBUG(("scsi%d: (%d:%d) selection timeout\n",
1583 host->host_no, pun, lun));
1584 NCR_700_scsi_done(hostdata, SCp, DID_NO_CONNECT<<16);
1585 } else if(sstat0 & PHASE_MISMATCH) {
1586 struct NCR_700_command_slot *slot = (SCp == NULL) ? NULL :
1587 (struct NCR_700_command_slot *)SCp->host_scribble;
1588
1589 if(dsp == Ent_SendMessage + 8 + hostdata->pScript) {
1590 /* It wants to reply to some part of
1591 * our message */
1592 #ifdef NCR_700_DEBUG
1593 __u32 temp = NCR_700_readl(host, TEMP_REG);
1594 int count = (hostdata->script[Ent_SendMessage/4] & 0xffffff) - ((NCR_700_readl(host, DBC_REG) & 0xffffff) + NCR_700_data_residual(host));
1595 printk("scsi%d (%d:%d) PHASE MISMATCH IN SEND MESSAGE %d remain, return %p[%04x], phase %s\n", host->host_no, pun, lun, count, (void *)temp, temp - hostdata->pScript, sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
1596 #endif
1597 resume_offset = hostdata->pScript + Ent_SendMessagePhaseMismatch;
1598 } else if(dsp >= to32bit(&slot->pSG[0].ins) &&
1599 dsp <= to32bit(&slot->pSG[NCR_700_SG_SEGMENTS].ins)) {
1600 int data_transfer = NCR_700_readl(host, DBC_REG) & 0xffffff;
1601 int SGcount = (dsp - to32bit(&slot->pSG[0].ins))/sizeof(struct NCR_700_SG_List);
1602 int residual = NCR_700_data_residual(host);
1603 int i;
1604 #ifdef NCR_700_DEBUG
1605 __u32 naddr = NCR_700_readl(host, DNAD_REG);
1606
1607 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x\n",
1608 host->host_no, pun, lun,
1609 SGcount, data_transfer);
1610 scsi_print_command(SCp);
1611 if(residual) {
1612 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x, residual %d\n",
1613 host->host_no, pun, lun,
1614 SGcount, data_transfer, residual);
1615 }
1616 #endif
1617 data_transfer += residual;
1618
1619 if(data_transfer != 0) {
1620 int count;
1621 __u32 pAddr;
1622
1623 SGcount--;
1624
1625 count = (bS_to_cpu(slot->SG[SGcount].ins) & 0x00ffffff);
1626 DEBUG(("DATA TRANSFER MISMATCH, count = %d, transferred %d\n", count, count-data_transfer));
1627 slot->SG[SGcount].ins &= bS_to_host(0xff000000);
1628 slot->SG[SGcount].ins |= bS_to_host(data_transfer);
1629 pAddr = bS_to_cpu(slot->SG[SGcount].pAddr);
1630 pAddr += (count - data_transfer);
1631 #ifdef NCR_700_DEBUG
1632 if(pAddr != naddr) {
1633 printk("scsi%d (%d:%d) transfer mismatch pAddr=%lx, naddr=%lx, data_transfer=%d, residual=%d\n", host->host_no, pun, lun, (unsigned long)pAddr, (unsigned long)naddr, data_transfer, residual);
1634 }
1635 #endif
1636 slot->SG[SGcount].pAddr = bS_to_host(pAddr);
1637 }
1638 /* set the executed moves to nops */
1639 for(i=0; i<SGcount; i++) {
1640 slot->SG[i].ins = bS_to_host(SCRIPT_NOP);
1641 slot->SG[i].pAddr = 0;
1642 }
1643 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1644 /* and pretend we disconnected after
1645 * the command phase */
1646 resume_offset = hostdata->pScript + Ent_MsgInDuringData;
1647 /* make sure all the data is flushed */
1648 NCR_700_flush_fifo(host);
1649 } else {
1650 __u8 sbcl = NCR_700_readb(host, SBCL_REG);
1651 printk(KERN_ERR "scsi%d: (%d:%d) phase mismatch at %04x, phase %s\n",
1652 host->host_no, pun, lun, dsp - hostdata->pScript, sbcl_to_string(sbcl));
1653 NCR_700_internal_bus_reset(host);
1654 }
1655
1656 } else if(sstat0 & SCSI_GROSS_ERROR) {
1657 printk(KERN_ERR "scsi%d: (%d:%d) GROSS ERROR\n",
1658 host->host_no, pun, lun);
1659 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1660 } else if(sstat0 & PARITY_ERROR) {
1661 printk(KERN_ERR "scsi%d: (%d:%d) PARITY ERROR\n",
1662 host->host_no, pun, lun);
1663 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1664 } else if(dstat & SCRIPT_INT_RECEIVED) {
1665 DEBUG(("scsi%d: (%d:%d) ====>SCRIPT INTERRUPT<====\n",
1666 host->host_no, pun, lun));
1667 resume_offset = process_script_interrupt(dsps, dsp, SCp, host, hostdata);
1668 } else if(dstat & (ILGL_INST_DETECTED)) {
1669 printk(KERN_ERR "scsi%d: (%d:%d) Illegal Instruction detected at 0x%08x[0x%x]!!!\n"
1670 " Please email James.Bottomley@HansenPartnership.com with the details\n",
1671 host->host_no, pun, lun,
1672 dsp, dsp - hostdata->pScript);
1673 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1674 } else if(dstat & (WATCH_DOG_INTERRUPT|ABORTED)) {
1675 printk(KERN_ERR "scsi%d: (%d:%d) serious DMA problem, dstat=%02x\n",
1676 host->host_no, pun, lun, dstat);
1677 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1678 }
1679
1680
1681 /* NOTE: selection interrupt processing MUST occur
1682 * after script interrupt processing to correctly cope
1683 * with the case where we process a disconnect and
1684 * then get reselected before we process the
1685 * disconnection */
1686 if(sstat0 & SELECTED) {
1687 /* FIXME: It currently takes at least FOUR
1688 * interrupts to complete a command that
1689 * disconnects: one for the disconnect, one
1690 * for the reselection, one to get the
1691 * reselection data and one to complete the
1692 * command. If we guess the reselected
1693 * command here and prepare it, we only need
1694 * to get a reselection data interrupt if we
1695 * guessed wrongly. Since the interrupt
1696 * overhead is much greater than the command
1697 * setup, this would be an efficient
1698 * optimisation particularly as we probably
1699 * only have one outstanding command on a
1700 * target most of the time */
1701
1702 resume_offset = process_selection(host, dsp);
1703
1704 }
1705
1706 }
1707
1708 if(resume_offset) {
1709 if(hostdata->state != NCR_700_HOST_BUSY) {
1710 printk(KERN_ERR "scsi%d: Driver error: resume at 0x%08x [0x%04x] with non busy host!\n",
1711 host->host_no, resume_offset, resume_offset - hostdata->pScript);
1712 hostdata->state = NCR_700_HOST_BUSY;
1713 }
1714
1715 DEBUG(("Attempting to resume at %x\n", resume_offset));
1716 NCR_700_clear_fifo(host);
1717 NCR_700_writel(resume_offset, host, DSP_REG);
1718 }
1719 /* There is probably a technical no-no about this: If we're a
1720 * shared interrupt and we got this interrupt because the
1721 * other device needs servicing not us, we're still going to
1722 * check our queued commands here---of course, there shouldn't
1723 * be any outstanding.... */
1724 if(hostdata->state == NCR_700_HOST_FREE) {
1725 int i;
1726
1727 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1728 /* fairness: always run the queue from the last
1729 * position we left off */
1730 int j = (i + hostdata->saved_slot_position)
1731 % NCR_700_COMMAND_SLOTS_PER_HOST;
1732
1733 if(hostdata->slots[j].state != NCR_700_SLOT_QUEUED)
1734 continue;
1735 if(NCR_700_start_command(hostdata->slots[j].cmnd)) {
1736 DEBUG(("scsi%d: Issuing saved command slot %p, cmd %p\t\n",
1737 host->host_no, &hostdata->slots[j],
1738 hostdata->slots[j].cmnd));
1739 hostdata->saved_slot_position = j + 1;
1740 }
1741
1742 break;
1743 }
1744 }
1745 out_unlock:
1746 spin_unlock_irqrestore(host->host_lock, flags);
1747 return IRQ_RETVAL(handled);
1748 }
1749
1750 static int
1751 NCR_700_queuecommand_lck(struct scsi_cmnd *SCp, void (*done)(struct scsi_cmnd *))
1752 {
1753 struct NCR_700_Host_Parameters *hostdata =
1754 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1755 __u32 move_ins;
1756 enum dma_data_direction direction;
1757 struct NCR_700_command_slot *slot;
1758
1759 if(hostdata->command_slot_count >= NCR_700_COMMAND_SLOTS_PER_HOST) {
1760 /* We're over our allocation, this should never happen
1761 * since we report the max allocation to the mid layer */
1762 printk(KERN_WARNING "scsi%d: Command depth has gone over queue depth\n", SCp->device->host->host_no);
1763 return 1;
1764 }
1765 /* check for untagged commands. We cannot have any outstanding
1766 * commands if we accept them. Commands could be untagged because:
1767 *
1768 * - The tag negotiated bitmap is clear
1769 * - The blk layer sent and untagged command
1770 */
1771 if(NCR_700_get_depth(SCp->device) != 0
1772 && (!(hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1773 || !(SCp->flags & SCMD_TAGGED))) {
1774 CDEBUG(KERN_ERR, SCp, "has non zero depth %d\n",
1775 NCR_700_get_depth(SCp->device));
1776 return SCSI_MLQUEUE_DEVICE_BUSY;
1777 }
1778 if(NCR_700_get_depth(SCp->device) >= SCp->device->queue_depth) {
1779 CDEBUG(KERN_ERR, SCp, "has max tag depth %d\n",
1780 NCR_700_get_depth(SCp->device));
1781 return SCSI_MLQUEUE_DEVICE_BUSY;
1782 }
1783 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) + 1);
1784
1785 /* begin the command here */
1786 /* no need to check for NULL, test for command_slot_count above
1787 * ensures a slot is free */
1788 slot = find_empty_slot(hostdata);
1789
1790 slot->cmnd = SCp;
1791
1792 SCp->scsi_done = done;
1793 SCp->host_scribble = (unsigned char *)slot;
1794 SCp->SCp.ptr = NULL;
1795 SCp->SCp.buffer = NULL;
1796
1797 #ifdef NCR_700_DEBUG
1798 printk("53c700: scsi%d, command ", SCp->device->host->host_no);
1799 scsi_print_command(SCp);
1800 #endif
1801 if ((SCp->flags & SCMD_TAGGED)
1802 && (hostdata->tag_negotiated &(1<<scmd_id(SCp))) == 0
1803 && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_START_TAG_NEGOTIATION) {
1804 scmd_printk(KERN_ERR, SCp, "Enabling Tag Command Queuing\n");
1805 hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1806 NCR_700_set_tag_neg_state(SCp->device, NCR_700_DURING_TAG_NEGOTIATION);
1807 }
1808
1809 /* here we may have to process an untagged command. The gate
1810 * above ensures that this will be the only one outstanding,
1811 * so clear the tag negotiated bit.
1812 *
1813 * FIXME: This will royally screw up on multiple LUN devices
1814 * */
1815 if (!(SCp->flags & SCMD_TAGGED)
1816 && (hostdata->tag_negotiated &(1<<scmd_id(SCp)))) {
1817 scmd_printk(KERN_INFO, SCp, "Disabling Tag Command Queuing\n");
1818 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1819 }
1820
1821 if((hostdata->tag_negotiated &(1<<scmd_id(SCp)))
1822 && scsi_get_tag_type(SCp->device)) {
1823 slot->tag = SCp->request->tag;
1824 CDEBUG(KERN_DEBUG, SCp, "sending out tag %d, slot %p\n",
1825 slot->tag, slot);
1826 } else {
1827 slot->tag = SCSI_NO_TAG;
1828 /* must populate current_cmnd for scsi_find_tag to work */
1829 SCp->device->current_cmnd = SCp;
1830 }
1831 /* sanity check: some of the commands generated by the mid-layer
1832 * have an eccentric idea of their sc_data_direction */
1833 if(!scsi_sg_count(SCp) && !scsi_bufflen(SCp) &&
1834 SCp->sc_data_direction != DMA_NONE) {
1835 #ifdef NCR_700_DEBUG
1836 printk("53c700: Command");
1837 scsi_print_command(SCp);
1838 printk("Has wrong data direction %d\n", SCp->sc_data_direction);
1839 #endif
1840 SCp->sc_data_direction = DMA_NONE;
1841 }
1842
1843 switch (SCp->cmnd[0]) {
1844 case REQUEST_SENSE:
1845 /* clear the internal sense magic */
1846 SCp->cmnd[6] = 0;
1847 /* fall through */
1848 default:
1849 /* OK, get it from the command */
1850 switch(SCp->sc_data_direction) {
1851 case DMA_BIDIRECTIONAL:
1852 default:
1853 printk(KERN_ERR "53c700: Unknown command for data direction ");
1854 scsi_print_command(SCp);
1855
1856 move_ins = 0;
1857 break;
1858 case DMA_NONE:
1859 move_ins = 0;
1860 break;
1861 case DMA_FROM_DEVICE:
1862 move_ins = SCRIPT_MOVE_DATA_IN;
1863 break;
1864 case DMA_TO_DEVICE:
1865 move_ins = SCRIPT_MOVE_DATA_OUT;
1866 break;
1867 }
1868 }
1869
1870 /* now build the scatter gather list */
1871 direction = SCp->sc_data_direction;
1872 if(move_ins != 0) {
1873 int i;
1874 int sg_count;
1875 dma_addr_t vPtr = 0;
1876 struct scatterlist *sg;
1877 __u32 count = 0;
1878
1879 sg_count = scsi_dma_map(SCp);
1880 BUG_ON(sg_count < 0);
1881
1882 scsi_for_each_sg(SCp, sg, sg_count, i) {
1883 vPtr = sg_dma_address(sg);
1884 count = sg_dma_len(sg);
1885
1886 slot->SG[i].ins = bS_to_host(move_ins | count);
1887 DEBUG((" scatter block %d: move %d[%08x] from 0x%lx\n",
1888 i, count, slot->SG[i].ins, (unsigned long)vPtr));
1889 slot->SG[i].pAddr = bS_to_host(vPtr);
1890 }
1891 slot->SG[i].ins = bS_to_host(SCRIPT_RETURN);
1892 slot->SG[i].pAddr = 0;
1893 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1894 DEBUG((" SETTING %08lx to %x\n",
1895 (&slot->pSG[i].ins),
1896 slot->SG[i].ins));
1897 }
1898 slot->resume_offset = 0;
1899 slot->pCmd = dma_map_single(hostdata->dev, SCp->cmnd,
1900 MAX_COMMAND_SIZE, DMA_TO_DEVICE);
1901 NCR_700_start_command(SCp);
1902 return 0;
1903 }
1904
1905 STATIC DEF_SCSI_QCMD(NCR_700_queuecommand)
1906
1907 STATIC int
1908 NCR_700_abort(struct scsi_cmnd * SCp)
1909 {
1910 struct NCR_700_command_slot *slot;
1911
1912 scmd_printk(KERN_INFO, SCp, "abort command\n");
1913
1914 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1915
1916 if(slot == NULL)
1917 /* no outstanding command to abort */
1918 return SUCCESS;
1919 if(SCp->cmnd[0] == TEST_UNIT_READY) {
1920 /* FIXME: This is because of a problem in the new
1921 * error handler. When it is in error recovery, it
1922 * will send a TUR to a device it thinks may still be
1923 * showing a problem. If the TUR isn't responded to,
1924 * it will abort it and mark the device off line.
1925 * Unfortunately, it does no other error recovery, so
1926 * this would leave us with an outstanding command
1927 * occupying a slot. Rather than allow this to
1928 * happen, we issue a bus reset to force all
1929 * outstanding commands to terminate here. */
1930 NCR_700_internal_bus_reset(SCp->device->host);
1931 /* still drop through and return failed */
1932 }
1933 return FAILED;
1934
1935 }
1936
1937 STATIC int
1938 NCR_700_bus_reset(struct scsi_cmnd * SCp)
1939 {
1940 DECLARE_COMPLETION_ONSTACK(complete);
1941 struct NCR_700_Host_Parameters *hostdata =
1942 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1943
1944 scmd_printk(KERN_INFO, SCp,
1945 "New error handler wants BUS reset, cmd %p\n\t", SCp);
1946 scsi_print_command(SCp);
1947
1948 /* In theory, eh_complete should always be null because the
1949 * eh is single threaded, but just in case we're handling a
1950 * reset via sg or something */
1951 spin_lock_irq(SCp->device->host->host_lock);
1952 while (hostdata->eh_complete != NULL) {
1953 spin_unlock_irq(SCp->device->host->host_lock);
1954 msleep_interruptible(100);
1955 spin_lock_irq(SCp->device->host->host_lock);
1956 }
1957
1958 hostdata->eh_complete = &complete;
1959 NCR_700_internal_bus_reset(SCp->device->host);
1960
1961 spin_unlock_irq(SCp->device->host->host_lock);
1962 wait_for_completion(&complete);
1963 spin_lock_irq(SCp->device->host->host_lock);
1964
1965 hostdata->eh_complete = NULL;
1966 /* Revalidate the transport parameters of the failing device */
1967 if(hostdata->fast)
1968 spi_schedule_dv_device(SCp->device);
1969
1970 spin_unlock_irq(SCp->device->host->host_lock);
1971 return SUCCESS;
1972 }
1973
1974 STATIC int
1975 NCR_700_host_reset(struct scsi_cmnd * SCp)
1976 {
1977 scmd_printk(KERN_INFO, SCp, "New error handler wants HOST reset\n\t");
1978 scsi_print_command(SCp);
1979
1980 spin_lock_irq(SCp->device->host->host_lock);
1981
1982 NCR_700_internal_bus_reset(SCp->device->host);
1983 NCR_700_chip_reset(SCp->device->host);
1984
1985 spin_unlock_irq(SCp->device->host->host_lock);
1986
1987 return SUCCESS;
1988 }
1989
1990 STATIC void
1991 NCR_700_set_period(struct scsi_target *STp, int period)
1992 {
1993 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
1994 struct NCR_700_Host_Parameters *hostdata =
1995 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
1996
1997 if(!hostdata->fast)
1998 return;
1999
2000 if(period < hostdata->min_period)
2001 period = hostdata->min_period;
2002
2003 spi_period(STp) = period;
2004 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2005 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2006 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2007 }
2008
2009 STATIC void
2010 NCR_700_set_offset(struct scsi_target *STp, int offset)
2011 {
2012 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
2013 struct NCR_700_Host_Parameters *hostdata =
2014 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
2015 int max_offset = hostdata->chip710
2016 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET;
2017
2018 if(!hostdata->fast)
2019 return;
2020
2021 if(offset > max_offset)
2022 offset = max_offset;
2023
2024 /* if we're currently async, make sure the period is reasonable */
2025 if(spi_offset(STp) == 0 && (spi_period(STp) < hostdata->min_period ||
2026 spi_period(STp) > 0xff))
2027 spi_period(STp) = hostdata->min_period;
2028
2029 spi_offset(STp) = offset;
2030 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2031 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2032 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2033 }
2034
2035 STATIC int
2036 NCR_700_slave_alloc(struct scsi_device *SDp)
2037 {
2038 SDp->hostdata = kzalloc(sizeof(struct NCR_700_Device_Parameters),
2039 GFP_KERNEL);
2040
2041 if (!SDp->hostdata)
2042 return -ENOMEM;
2043
2044 return 0;
2045 }
2046
2047 STATIC int
2048 NCR_700_slave_configure(struct scsi_device *SDp)
2049 {
2050 struct NCR_700_Host_Parameters *hostdata =
2051 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2052
2053 /* to do here: allocate memory; build a queue_full list */
2054 if(SDp->tagged_supported) {
2055 scsi_change_queue_depth(SDp, NCR_700_DEFAULT_TAGS);
2056 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2057 }
2058
2059 if(hostdata->fast) {
2060 /* Find the correct offset and period via domain validation */
2061 if (!spi_initial_dv(SDp->sdev_target))
2062 spi_dv_device(SDp);
2063 } else {
2064 spi_offset(SDp->sdev_target) = 0;
2065 spi_period(SDp->sdev_target) = 0;
2066 }
2067 return 0;
2068 }
2069
2070 STATIC void
2071 NCR_700_slave_destroy(struct scsi_device *SDp)
2072 {
2073 kfree(SDp->hostdata);
2074 SDp->hostdata = NULL;
2075 }
2076
2077 static int
2078 NCR_700_change_queue_depth(struct scsi_device *SDp, int depth)
2079 {
2080 if (depth > NCR_700_MAX_TAGS)
2081 depth = NCR_700_MAX_TAGS;
2082 return scsi_change_queue_depth(SDp, depth);
2083 }
2084
2085 static int NCR_700_change_queue_type(struct scsi_device *SDp, int tag_type)
2086 {
2087 int change_tag = ((tag_type ==0 && scsi_get_tag_type(SDp) != 0)
2088 || (tag_type != 0 && scsi_get_tag_type(SDp) == 0));
2089 struct NCR_700_Host_Parameters *hostdata =
2090 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2091
2092 /* We have a global (per target) flag to track whether TCQ is
2093 * enabled, so we'll be turning it off for the entire target here.
2094 * our tag algorithm will fail if we mix tagged and untagged commands,
2095 * so quiesce the device before doing this */
2096 if (change_tag)
2097 scsi_target_quiesce(SDp->sdev_target);
2098
2099 scsi_set_tag_type(SDp, tag_type);
2100 if (!tag_type) {
2101 /* shift back to the default unqueued number of commands
2102 * (the user can still raise this) */
2103 scsi_change_queue_depth(SDp, SDp->host->cmd_per_lun);
2104 hostdata->tag_negotiated &= ~(1 << sdev_id(SDp));
2105 } else {
2106 /* Here, we cleared the negotiation flag above, so this
2107 * will force the driver to renegotiate */
2108 scsi_change_queue_depth(SDp, SDp->queue_depth);
2109 if (change_tag)
2110 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2111 }
2112 if (change_tag)
2113 scsi_target_resume(SDp->sdev_target);
2114
2115 return tag_type;
2116 }
2117
2118 static ssize_t
2119 NCR_700_show_active_tags(struct device *dev, struct device_attribute *attr, char *buf)
2120 {
2121 struct scsi_device *SDp = to_scsi_device(dev);
2122
2123 return snprintf(buf, 20, "%d\n", NCR_700_get_depth(SDp));
2124 }
2125
2126 static struct device_attribute NCR_700_active_tags_attr = {
2127 .attr = {
2128 .name = "active_tags",
2129 .mode = S_IRUGO,
2130 },
2131 .show = NCR_700_show_active_tags,
2132 };
2133
2134 STATIC struct device_attribute *NCR_700_dev_attrs[] = {
2135 &NCR_700_active_tags_attr,
2136 NULL,
2137 };
2138
2139 EXPORT_SYMBOL(NCR_700_detect);
2140 EXPORT_SYMBOL(NCR_700_release);
2141 EXPORT_SYMBOL(NCR_700_intr);
2142
2143 static struct spi_function_template NCR_700_transport_functions = {
2144 .set_period = NCR_700_set_period,
2145 .show_period = 1,
2146 .set_offset = NCR_700_set_offset,
2147 .show_offset = 1,
2148 };
2149
2150 static int __init NCR_700_init(void)
2151 {
2152 NCR_700_transport_template = spi_attach_transport(&NCR_700_transport_functions);
2153 if(!NCR_700_transport_template)
2154 return -ENODEV;
2155 return 0;
2156 }
2157
2158 static void __exit NCR_700_exit(void)
2159 {
2160 spi_release_transport(NCR_700_transport_template);
2161 }
2162
2163 module_init(NCR_700_init);
2164 module_exit(NCR_700_exit);
2165