]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/blob - drivers/scsi/53c700.c
Remove obsolete #include <linux/config.h>
[mirror_ubuntu-jammy-kernel.git] / drivers / scsi / 53c700.c
1 /* -*- mode: c; c-basic-offset: 8 -*- */
2
3 /* NCR (or Symbios) 53c700 and 53c700-66 Driver
4 *
5 * Copyright (C) 2001 by James.Bottomley@HansenPartnership.com
6 **-----------------------------------------------------------------------------
7 **
8 ** This program is free software; you can redistribute it and/or modify
9 ** it under the terms of the GNU General Public License as published by
10 ** the Free Software Foundation; either version 2 of the License, or
11 ** (at your option) any later version.
12 **
13 ** This program is distributed in the hope that it will be useful,
14 ** but WITHOUT ANY WARRANTY; without even the implied warranty of
15 ** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 ** GNU General Public License for more details.
17 **
18 ** You should have received a copy of the GNU General Public License
19 ** along with this program; if not, write to the Free Software
20 ** Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
21 **
22 **-----------------------------------------------------------------------------
23 */
24
25 /* Notes:
26 *
27 * This driver is designed exclusively for these chips (virtually the
28 * earliest of the scripts engine chips). They need their own drivers
29 * because they are missing so many of the scripts and snazzy register
30 * features of their elder brothers (the 710, 720 and 770).
31 *
32 * The 700 is the lowliest of the line, it can only do async SCSI.
33 * The 700-66 can at least do synchronous SCSI up to 10MHz.
34 *
35 * The 700 chip has no host bus interface logic of its own. However,
36 * it is usually mapped to a location with well defined register
37 * offsets. Therefore, if you can determine the base address and the
38 * irq your board incorporating this chip uses, you can probably use
39 * this driver to run it (although you'll probably have to write a
40 * minimal wrapper for the purpose---see the NCR_D700 driver for
41 * details about how to do this).
42 *
43 *
44 * TODO List:
45 *
46 * 1. Better statistics in the proc fs
47 *
48 * 2. Implement message queue (queues SCSI messages like commands) and make
49 * the abort and device reset functions use them.
50 * */
51
52 /* CHANGELOG
53 *
54 * Version 2.8
55 *
56 * Fixed bad bug affecting tag starvation processing (previously the
57 * driver would hang the system if too many tags starved. Also fixed
58 * bad bug having to do with 10 byte command processing and REQUEST
59 * SENSE (the command would loop forever getting a transfer length
60 * mismatch in the CMD phase).
61 *
62 * Version 2.7
63 *
64 * Fixed scripts problem which caused certain devices (notably CDRWs)
65 * to hang on initial INQUIRY. Updated NCR_700_readl/writel to use
66 * __raw_readl/writel for parisc compatibility (Thomas
67 * Bogendoerfer). Added missing SCp->request_bufflen initialisation
68 * for sense requests (Ryan Bradetich).
69 *
70 * Version 2.6
71 *
72 * Following test of the 64 bit parisc kernel by Richard Hirst,
73 * several problems have now been corrected. Also adds support for
74 * consistent memory allocation.
75 *
76 * Version 2.5
77 *
78 * More Compatibility changes for 710 (now actually works). Enhanced
79 * support for odd clock speeds which constrain SDTR negotiations.
80 * correct cacheline separation for scsi messages and status for
81 * incoherent architectures. Use of the pci mapping functions on
82 * buffers to begin support for 64 bit drivers.
83 *
84 * Version 2.4
85 *
86 * Added support for the 53c710 chip (in 53c700 emulation mode only---no
87 * special 53c710 instructions or registers are used).
88 *
89 * Version 2.3
90 *
91 * More endianness/cache coherency changes.
92 *
93 * Better bad device handling (handles devices lying about tag
94 * queueing support and devices which fail to provide sense data on
95 * contingent allegiance conditions)
96 *
97 * Many thanks to Richard Hirst <rhirst@linuxcare.com> for patiently
98 * debugging this driver on the parisc architecture and suggesting
99 * many improvements and bug fixes.
100 *
101 * Thanks also go to Linuxcare Inc. for providing several PARISC
102 * machines for me to debug the driver on.
103 *
104 * Version 2.2
105 *
106 * Made the driver mem or io mapped; added endian invariance; added
107 * dma cache flushing operations for architectures which need it;
108 * added support for more varied clocking speeds.
109 *
110 * Version 2.1
111 *
112 * Initial modularisation from the D700. See NCR_D700.c for the rest of
113 * the changelog.
114 * */
115 #define NCR_700_VERSION "2.8"
116
117 #include <linux/kernel.h>
118 #include <linux/types.h>
119 #include <linux/string.h>
120 #include <linux/ioport.h>
121 #include <linux/delay.h>
122 #include <linux/spinlock.h>
123 #include <linux/completion.h>
124 #include <linux/sched.h>
125 #include <linux/init.h>
126 #include <linux/proc_fs.h>
127 #include <linux/blkdev.h>
128 #include <linux/module.h>
129 #include <linux/interrupt.h>
130 #include <linux/device.h>
131 #include <asm/dma.h>
132 #include <asm/system.h>
133 #include <asm/io.h>
134 #include <asm/pgtable.h>
135 #include <asm/byteorder.h>
136
137 #include <scsi/scsi.h>
138 #include <scsi/scsi_cmnd.h>
139 #include <scsi/scsi_dbg.h>
140 #include <scsi/scsi_eh.h>
141 #include <scsi/scsi_host.h>
142 #include <scsi/scsi_tcq.h>
143 #include <scsi/scsi_transport.h>
144 #include <scsi/scsi_transport_spi.h>
145
146 #include "53c700.h"
147
148 /* NOTE: For 64 bit drivers there are points in the code where we use
149 * a non dereferenceable pointer to point to a structure in dma-able
150 * memory (which is 32 bits) so that we can use all of the structure
151 * operations but take the address at the end. This macro allows us
152 * to truncate the 64 bit pointer down to 32 bits without the compiler
153 * complaining */
154 #define to32bit(x) ((__u32)((unsigned long)(x)))
155
156 #ifdef NCR_700_DEBUG
157 #define STATIC
158 #else
159 #define STATIC static
160 #endif
161
162 MODULE_AUTHOR("James Bottomley");
163 MODULE_DESCRIPTION("53c700 and 53c700-66 Driver");
164 MODULE_LICENSE("GPL");
165
166 /* This is the script */
167 #include "53c700_d.h"
168
169
170 STATIC int NCR_700_queuecommand(struct scsi_cmnd *, void (*done)(struct scsi_cmnd *));
171 STATIC int NCR_700_abort(struct scsi_cmnd * SCpnt);
172 STATIC int NCR_700_bus_reset(struct scsi_cmnd * SCpnt);
173 STATIC int NCR_700_host_reset(struct scsi_cmnd * SCpnt);
174 STATIC void NCR_700_chip_setup(struct Scsi_Host *host);
175 STATIC void NCR_700_chip_reset(struct Scsi_Host *host);
176 STATIC int NCR_700_slave_configure(struct scsi_device *SDpnt);
177 STATIC void NCR_700_slave_destroy(struct scsi_device *SDpnt);
178 static int NCR_700_change_queue_depth(struct scsi_device *SDpnt, int depth);
179 static int NCR_700_change_queue_type(struct scsi_device *SDpnt, int depth);
180
181 STATIC struct device_attribute *NCR_700_dev_attrs[];
182
183 STATIC struct scsi_transport_template *NCR_700_transport_template = NULL;
184
185 struct NCR_700_sense {
186 unsigned char cmnd[MAX_COMMAND_SIZE];
187 };
188
189 static char *NCR_700_phase[] = {
190 "",
191 "after selection",
192 "before command phase",
193 "after command phase",
194 "after status phase",
195 "after data in phase",
196 "after data out phase",
197 "during data phase",
198 };
199
200 static char *NCR_700_condition[] = {
201 "",
202 "NOT MSG_OUT",
203 "UNEXPECTED PHASE",
204 "NOT MSG_IN",
205 "UNEXPECTED MSG",
206 "MSG_IN",
207 "SDTR_MSG RECEIVED",
208 "REJECT_MSG RECEIVED",
209 "DISCONNECT_MSG RECEIVED",
210 "MSG_OUT",
211 "DATA_IN",
212
213 };
214
215 static char *NCR_700_fatal_messages[] = {
216 "unexpected message after reselection",
217 "still MSG_OUT after message injection",
218 "not MSG_IN after selection",
219 "Illegal message length received",
220 };
221
222 static char *NCR_700_SBCL_bits[] = {
223 "IO ",
224 "CD ",
225 "MSG ",
226 "ATN ",
227 "SEL ",
228 "BSY ",
229 "ACK ",
230 "REQ ",
231 };
232
233 static char *NCR_700_SBCL_to_phase[] = {
234 "DATA_OUT",
235 "DATA_IN",
236 "CMD_OUT",
237 "STATE",
238 "ILLEGAL PHASE",
239 "ILLEGAL PHASE",
240 "MSG OUT",
241 "MSG IN",
242 };
243
244 /* This translates the SDTR message offset and period to a value
245 * which can be loaded into the SXFER_REG.
246 *
247 * NOTE: According to SCSI-2, the true transfer period (in ns) is
248 * actually four times this period value */
249 static inline __u8
250 NCR_700_offset_period_to_sxfer(struct NCR_700_Host_Parameters *hostdata,
251 __u8 offset, __u8 period)
252 {
253 int XFERP;
254
255 __u8 min_xferp = (hostdata->chip710
256 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
257 __u8 max_offset = (hostdata->chip710
258 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET);
259
260 if(offset == 0)
261 return 0;
262
263 if(period < hostdata->min_period) {
264 printk(KERN_WARNING "53c700: Period %dns is less than this chip's minimum, setting to %d\n", period*4, NCR_700_MIN_PERIOD*4);
265 period = hostdata->min_period;
266 }
267 XFERP = (period*4 * hostdata->sync_clock)/1000 - 4;
268 if(offset > max_offset) {
269 printk(KERN_WARNING "53c700: Offset %d exceeds chip maximum, setting to %d\n",
270 offset, max_offset);
271 offset = max_offset;
272 }
273 if(XFERP < min_xferp) {
274 printk(KERN_WARNING "53c700: XFERP %d is less than minium, setting to %d\n",
275 XFERP, min_xferp);
276 XFERP = min_xferp;
277 }
278 return (offset & 0x0f) | (XFERP & 0x07)<<4;
279 }
280
281 static inline __u8
282 NCR_700_get_SXFER(struct scsi_device *SDp)
283 {
284 struct NCR_700_Host_Parameters *hostdata =
285 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
286
287 return NCR_700_offset_period_to_sxfer(hostdata,
288 spi_offset(SDp->sdev_target),
289 spi_period(SDp->sdev_target));
290 }
291
292 struct Scsi_Host *
293 NCR_700_detect(struct scsi_host_template *tpnt,
294 struct NCR_700_Host_Parameters *hostdata, struct device *dev)
295 {
296 dma_addr_t pScript, pSlots;
297 __u8 *memory;
298 __u32 *script;
299 struct Scsi_Host *host;
300 static int banner = 0;
301 int j;
302
303 if(tpnt->sdev_attrs == NULL)
304 tpnt->sdev_attrs = NCR_700_dev_attrs;
305
306 memory = dma_alloc_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
307 &pScript, GFP_KERNEL);
308 if(memory == NULL) {
309 printk(KERN_ERR "53c700: Failed to allocate memory for driver, detatching\n");
310 return NULL;
311 }
312
313 script = (__u32 *)memory;
314 hostdata->msgin = memory + MSGIN_OFFSET;
315 hostdata->msgout = memory + MSGOUT_OFFSET;
316 hostdata->status = memory + STATUS_OFFSET;
317 /* all of these offsets are L1_CACHE_BYTES separated. It is fatal
318 * if this isn't sufficient separation to avoid dma flushing issues */
319 BUG_ON(!dma_is_consistent(pScript) && L1_CACHE_BYTES < dma_get_cache_alignment());
320 hostdata->slots = (struct NCR_700_command_slot *)(memory + SLOTS_OFFSET);
321 hostdata->dev = dev;
322
323 pSlots = pScript + SLOTS_OFFSET;
324
325 /* Fill in the missing routines from the host template */
326 tpnt->queuecommand = NCR_700_queuecommand;
327 tpnt->eh_abort_handler = NCR_700_abort;
328 tpnt->eh_bus_reset_handler = NCR_700_bus_reset;
329 tpnt->eh_host_reset_handler = NCR_700_host_reset;
330 tpnt->can_queue = NCR_700_COMMAND_SLOTS_PER_HOST;
331 tpnt->sg_tablesize = NCR_700_SG_SEGMENTS;
332 tpnt->cmd_per_lun = NCR_700_CMD_PER_LUN;
333 tpnt->use_clustering = ENABLE_CLUSTERING;
334 tpnt->slave_configure = NCR_700_slave_configure;
335 tpnt->slave_destroy = NCR_700_slave_destroy;
336 tpnt->change_queue_depth = NCR_700_change_queue_depth;
337 tpnt->change_queue_type = NCR_700_change_queue_type;
338
339 if(tpnt->name == NULL)
340 tpnt->name = "53c700";
341 if(tpnt->proc_name == NULL)
342 tpnt->proc_name = "53c700";
343
344 host = scsi_host_alloc(tpnt, 4);
345 if (!host)
346 return NULL;
347 memset(hostdata->slots, 0, sizeof(struct NCR_700_command_slot)
348 * NCR_700_COMMAND_SLOTS_PER_HOST);
349 for (j = 0; j < NCR_700_COMMAND_SLOTS_PER_HOST; j++) {
350 dma_addr_t offset = (dma_addr_t)((unsigned long)&hostdata->slots[j].SG[0]
351 - (unsigned long)&hostdata->slots[0].SG[0]);
352 hostdata->slots[j].pSG = (struct NCR_700_SG_List *)((unsigned long)(pSlots + offset));
353 if(j == 0)
354 hostdata->free_list = &hostdata->slots[j];
355 else
356 hostdata->slots[j-1].ITL_forw = &hostdata->slots[j];
357 hostdata->slots[j].state = NCR_700_SLOT_FREE;
358 }
359
360 for (j = 0; j < ARRAY_SIZE(SCRIPT); j++)
361 script[j] = bS_to_host(SCRIPT[j]);
362
363 /* adjust all labels to be bus physical */
364 for (j = 0; j < PATCHES; j++)
365 script[LABELPATCHES[j]] = bS_to_host(pScript + SCRIPT[LABELPATCHES[j]]);
366 /* now patch up fixed addresses. */
367 script_patch_32(script, MessageLocation,
368 pScript + MSGOUT_OFFSET);
369 script_patch_32(script, StatusAddress,
370 pScript + STATUS_OFFSET);
371 script_patch_32(script, ReceiveMsgAddress,
372 pScript + MSGIN_OFFSET);
373
374 hostdata->script = script;
375 hostdata->pScript = pScript;
376 dma_sync_single_for_device(hostdata->dev, pScript, sizeof(SCRIPT), DMA_TO_DEVICE);
377 hostdata->state = NCR_700_HOST_FREE;
378 hostdata->cmd = NULL;
379 host->max_id = 8;
380 host->max_lun = NCR_700_MAX_LUNS;
381 BUG_ON(NCR_700_transport_template == NULL);
382 host->transportt = NCR_700_transport_template;
383 host->unique_id = (unsigned long)hostdata->base;
384 hostdata->eh_complete = NULL;
385 host->hostdata[0] = (unsigned long)hostdata;
386 /* kick the chip */
387 NCR_700_writeb(0xff, host, CTEST9_REG);
388 if (hostdata->chip710)
389 hostdata->rev = (NCR_700_readb(host, CTEST8_REG)>>4) & 0x0f;
390 else
391 hostdata->rev = (NCR_700_readb(host, CTEST7_REG)>>4) & 0x0f;
392 hostdata->fast = (NCR_700_readb(host, CTEST9_REG) == 0);
393 if (banner == 0) {
394 printk(KERN_NOTICE "53c700: Version " NCR_700_VERSION " By James.Bottomley@HansenPartnership.com\n");
395 banner = 1;
396 }
397 printk(KERN_NOTICE "scsi%d: %s rev %d %s\n", host->host_no,
398 hostdata->chip710 ? "53c710" :
399 (hostdata->fast ? "53c700-66" : "53c700"),
400 hostdata->rev, hostdata->differential ?
401 "(Differential)" : "");
402 /* reset the chip */
403 NCR_700_chip_reset(host);
404
405 if (scsi_add_host(host, dev)) {
406 dev_printk(KERN_ERR, dev, "53c700: scsi_add_host failed\n");
407 scsi_host_put(host);
408 return NULL;
409 }
410
411 spi_signalling(host) = hostdata->differential ? SPI_SIGNAL_HVD :
412 SPI_SIGNAL_SE;
413
414 return host;
415 }
416
417 int
418 NCR_700_release(struct Scsi_Host *host)
419 {
420 struct NCR_700_Host_Parameters *hostdata =
421 (struct NCR_700_Host_Parameters *)host->hostdata[0];
422
423 dma_free_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
424 hostdata->script, hostdata->pScript);
425 return 1;
426 }
427
428 static inline __u8
429 NCR_700_identify(int can_disconnect, __u8 lun)
430 {
431 return IDENTIFY_BASE |
432 ((can_disconnect) ? 0x40 : 0) |
433 (lun & NCR_700_LUN_MASK);
434 }
435
436 /*
437 * Function : static int data_residual (Scsi_Host *host)
438 *
439 * Purpose : return residual data count of what's in the chip. If you
440 * really want to know what this function is doing, it's almost a
441 * direct transcription of the algorithm described in the 53c710
442 * guide, except that the DBC and DFIFO registers are only 6 bits
443 * wide on a 53c700.
444 *
445 * Inputs : host - SCSI host */
446 static inline int
447 NCR_700_data_residual (struct Scsi_Host *host) {
448 struct NCR_700_Host_Parameters *hostdata =
449 (struct NCR_700_Host_Parameters *)host->hostdata[0];
450 int count, synchronous = 0;
451 unsigned int ddir;
452
453 if(hostdata->chip710) {
454 count = ((NCR_700_readb(host, DFIFO_REG) & 0x7f) -
455 (NCR_700_readl(host, DBC_REG) & 0x7f)) & 0x7f;
456 } else {
457 count = ((NCR_700_readb(host, DFIFO_REG) & 0x3f) -
458 (NCR_700_readl(host, DBC_REG) & 0x3f)) & 0x3f;
459 }
460
461 if(hostdata->fast)
462 synchronous = NCR_700_readb(host, SXFER_REG) & 0x0f;
463
464 /* get the data direction */
465 ddir = NCR_700_readb(host, CTEST0_REG) & 0x01;
466
467 if (ddir) {
468 /* Receive */
469 if (synchronous)
470 count += (NCR_700_readb(host, SSTAT2_REG) & 0xf0) >> 4;
471 else
472 if (NCR_700_readb(host, SSTAT1_REG) & SIDL_REG_FULL)
473 ++count;
474 } else {
475 /* Send */
476 __u8 sstat = NCR_700_readb(host, SSTAT1_REG);
477 if (sstat & SODL_REG_FULL)
478 ++count;
479 if (synchronous && (sstat & SODR_REG_FULL))
480 ++count;
481 }
482 #ifdef NCR_700_DEBUG
483 if(count)
484 printk("RESIDUAL IS %d (ddir %d)\n", count, ddir);
485 #endif
486 return count;
487 }
488
489 /* print out the SCSI wires and corresponding phase from the SBCL register
490 * in the chip */
491 static inline char *
492 sbcl_to_string(__u8 sbcl)
493 {
494 int i;
495 static char ret[256];
496
497 ret[0]='\0';
498 for(i=0; i<8; i++) {
499 if((1<<i) & sbcl)
500 strcat(ret, NCR_700_SBCL_bits[i]);
501 }
502 strcat(ret, NCR_700_SBCL_to_phase[sbcl & 0x07]);
503 return ret;
504 }
505
506 static inline __u8
507 bitmap_to_number(__u8 bitmap)
508 {
509 __u8 i;
510
511 for(i=0; i<8 && !(bitmap &(1<<i)); i++)
512 ;
513 return i;
514 }
515
516 /* Pull a slot off the free list */
517 STATIC struct NCR_700_command_slot *
518 find_empty_slot(struct NCR_700_Host_Parameters *hostdata)
519 {
520 struct NCR_700_command_slot *slot = hostdata->free_list;
521
522 if(slot == NULL) {
523 /* sanity check */
524 if(hostdata->command_slot_count != NCR_700_COMMAND_SLOTS_PER_HOST)
525 printk(KERN_ERR "SLOTS FULL, but count is %d, should be %d\n", hostdata->command_slot_count, NCR_700_COMMAND_SLOTS_PER_HOST);
526 return NULL;
527 }
528
529 if(slot->state != NCR_700_SLOT_FREE)
530 /* should panic! */
531 printk(KERN_ERR "BUSY SLOT ON FREE LIST!!!\n");
532
533
534 hostdata->free_list = slot->ITL_forw;
535 slot->ITL_forw = NULL;
536
537
538 /* NOTE: set the state to busy here, not queued, since this
539 * indicates the slot is in use and cannot be run by the IRQ
540 * finish routine. If we cannot queue the command when it
541 * is properly build, we then change to NCR_700_SLOT_QUEUED */
542 slot->state = NCR_700_SLOT_BUSY;
543 slot->flags = 0;
544 hostdata->command_slot_count++;
545
546 return slot;
547 }
548
549 STATIC void
550 free_slot(struct NCR_700_command_slot *slot,
551 struct NCR_700_Host_Parameters *hostdata)
552 {
553 if((slot->state & NCR_700_SLOT_MASK) != NCR_700_SLOT_MAGIC) {
554 printk(KERN_ERR "53c700: SLOT %p is not MAGIC!!!\n", slot);
555 }
556 if(slot->state == NCR_700_SLOT_FREE) {
557 printk(KERN_ERR "53c700: SLOT %p is FREE!!!\n", slot);
558 }
559
560 slot->resume_offset = 0;
561 slot->cmnd = NULL;
562 slot->state = NCR_700_SLOT_FREE;
563 slot->ITL_forw = hostdata->free_list;
564 hostdata->free_list = slot;
565 hostdata->command_slot_count--;
566 }
567
568
569 /* This routine really does very little. The command is indexed on
570 the ITL and (if tagged) the ITLQ lists in _queuecommand */
571 STATIC void
572 save_for_reselection(struct NCR_700_Host_Parameters *hostdata,
573 struct scsi_cmnd *SCp, __u32 dsp)
574 {
575 /* Its just possible that this gets executed twice */
576 if(SCp != NULL) {
577 struct NCR_700_command_slot *slot =
578 (struct NCR_700_command_slot *)SCp->host_scribble;
579
580 slot->resume_offset = dsp;
581 }
582 hostdata->state = NCR_700_HOST_FREE;
583 hostdata->cmd = NULL;
584 }
585
586 STATIC inline void
587 NCR_700_unmap(struct NCR_700_Host_Parameters *hostdata, struct scsi_cmnd *SCp,
588 struct NCR_700_command_slot *slot)
589 {
590 if(SCp->sc_data_direction != DMA_NONE &&
591 SCp->sc_data_direction != DMA_BIDIRECTIONAL) {
592 if(SCp->use_sg) {
593 dma_unmap_sg(hostdata->dev, SCp->request_buffer,
594 SCp->use_sg, SCp->sc_data_direction);
595 } else {
596 dma_unmap_single(hostdata->dev, slot->dma_handle,
597 SCp->request_bufflen,
598 SCp->sc_data_direction);
599 }
600 }
601 }
602
603 STATIC inline void
604 NCR_700_scsi_done(struct NCR_700_Host_Parameters *hostdata,
605 struct scsi_cmnd *SCp, int result)
606 {
607 hostdata->state = NCR_700_HOST_FREE;
608 hostdata->cmd = NULL;
609
610 if(SCp != NULL) {
611 struct NCR_700_command_slot *slot =
612 (struct NCR_700_command_slot *)SCp->host_scribble;
613
614 NCR_700_unmap(hostdata, SCp, slot);
615 if (slot->flags == NCR_700_FLAG_AUTOSENSE) {
616 struct NCR_700_sense *sense = SCp->device->hostdata;
617 #ifdef NCR_700_DEBUG
618 printk(" ORIGINAL CMD %p RETURNED %d, new return is %d sense is\n",
619 SCp, SCp->cmnd[7], result);
620 scsi_print_sense("53c700", SCp);
621
622 #endif
623 dma_unmap_single(hostdata->dev, slot->dma_handle, sizeof(SCp->sense_buffer), DMA_FROM_DEVICE);
624 /* restore the old result if the request sense was
625 * successful */
626 if(result == 0)
627 result = sense->cmnd[7];
628 } else
629 dma_unmap_single(hostdata->dev, slot->pCmd,
630 sizeof(SCp->cmnd), DMA_TO_DEVICE);
631
632 free_slot(slot, hostdata);
633 #ifdef NCR_700_DEBUG
634 if(NCR_700_get_depth(SCp->device) == 0 ||
635 NCR_700_get_depth(SCp->device) > SCp->device->queue_depth)
636 printk(KERN_ERR "Invalid depth in NCR_700_scsi_done(): %d\n",
637 NCR_700_get_depth(SCp->device));
638 #endif /* NCR_700_DEBUG */
639 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) - 1);
640
641 SCp->host_scribble = NULL;
642 SCp->result = result;
643 SCp->scsi_done(SCp);
644 } else {
645 printk(KERN_ERR "53c700: SCSI DONE HAS NULL SCp\n");
646 }
647 }
648
649
650 STATIC void
651 NCR_700_internal_bus_reset(struct Scsi_Host *host)
652 {
653 /* Bus reset */
654 NCR_700_writeb(ASSERT_RST, host, SCNTL1_REG);
655 udelay(50);
656 NCR_700_writeb(0, host, SCNTL1_REG);
657
658 }
659
660 STATIC void
661 NCR_700_chip_setup(struct Scsi_Host *host)
662 {
663 struct NCR_700_Host_Parameters *hostdata =
664 (struct NCR_700_Host_Parameters *)host->hostdata[0];
665 __u32 dcntl_extra = 0;
666 __u8 min_period;
667 __u8 min_xferp = (hostdata->chip710 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
668
669 if(hostdata->chip710) {
670 __u8 burst_disable = hostdata->burst_disable
671 ? BURST_DISABLE : 0;
672 dcntl_extra = COMPAT_700_MODE;
673
674 NCR_700_writeb(dcntl_extra, host, DCNTL_REG);
675 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra,
676 host, DMODE_710_REG);
677 NCR_700_writeb(burst_disable | (hostdata->differential ?
678 DIFF : 0), host, CTEST7_REG);
679 NCR_700_writeb(BTB_TIMER_DISABLE, host, CTEST0_REG);
680 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY | PARITY
681 | AUTO_ATN, host, SCNTL0_REG);
682 } else {
683 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra,
684 host, DMODE_700_REG);
685 NCR_700_writeb(hostdata->differential ?
686 DIFF : 0, host, CTEST7_REG);
687 if(hostdata->fast) {
688 /* this is for 700-66, does nothing on 700 */
689 NCR_700_writeb(LAST_DIS_ENBL | ENABLE_ACTIVE_NEGATION
690 | GENERATE_RECEIVE_PARITY, host,
691 CTEST8_REG);
692 } else {
693 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY
694 | PARITY | AUTO_ATN, host, SCNTL0_REG);
695 }
696 }
697
698 NCR_700_writeb(1 << host->this_id, host, SCID_REG);
699 NCR_700_writeb(0, host, SBCL_REG);
700 NCR_700_writeb(ASYNC_OPERATION, host, SXFER_REG);
701
702 NCR_700_writeb(PHASE_MM_INT | SEL_TIMEOUT_INT | GROSS_ERR_INT | UX_DISC_INT
703 | RST_INT | PAR_ERR_INT | SELECT_INT, host, SIEN_REG);
704
705 NCR_700_writeb(ABORT_INT | INT_INST_INT | ILGL_INST_INT, host, DIEN_REG);
706 NCR_700_writeb(ENABLE_SELECT, host, SCNTL1_REG);
707 if(hostdata->clock > 75) {
708 printk(KERN_ERR "53c700: Clock speed %dMHz is too high: 75Mhz is the maximum this chip can be driven at\n", hostdata->clock);
709 /* do the best we can, but the async clock will be out
710 * of spec: sync divider 2, async divider 3 */
711 DEBUG(("53c700: sync 2 async 3\n"));
712 NCR_700_writeb(SYNC_DIV_2_0, host, SBCL_REG);
713 NCR_700_writeb(ASYNC_DIV_3_0 | dcntl_extra, host, DCNTL_REG);
714 hostdata->sync_clock = hostdata->clock/2;
715 } else if(hostdata->clock > 50 && hostdata->clock <= 75) {
716 /* sync divider 1.5, async divider 3 */
717 DEBUG(("53c700: sync 1.5 async 3\n"));
718 NCR_700_writeb(SYNC_DIV_1_5, host, SBCL_REG);
719 NCR_700_writeb(ASYNC_DIV_3_0 | dcntl_extra, host, DCNTL_REG);
720 hostdata->sync_clock = hostdata->clock*2;
721 hostdata->sync_clock /= 3;
722
723 } else if(hostdata->clock > 37 && hostdata->clock <= 50) {
724 /* sync divider 1, async divider 2 */
725 DEBUG(("53c700: sync 1 async 2\n"));
726 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
727 NCR_700_writeb(ASYNC_DIV_2_0 | dcntl_extra, host, DCNTL_REG);
728 hostdata->sync_clock = hostdata->clock;
729 } else if(hostdata->clock > 25 && hostdata->clock <=37) {
730 /* sync divider 1, async divider 1.5 */
731 DEBUG(("53c700: sync 1 async 1.5\n"));
732 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
733 NCR_700_writeb(ASYNC_DIV_1_5 | dcntl_extra, host, DCNTL_REG);
734 hostdata->sync_clock = hostdata->clock;
735 } else {
736 DEBUG(("53c700: sync 1 async 1\n"));
737 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
738 NCR_700_writeb(ASYNC_DIV_1_0 | dcntl_extra, host, DCNTL_REG);
739 /* sync divider 1, async divider 1 */
740 hostdata->sync_clock = hostdata->clock;
741 }
742 /* Calculate the actual minimum period that can be supported
743 * by our synchronous clock speed. See the 710 manual for
744 * exact details of this calculation which is based on a
745 * setting of the SXFER register */
746 min_period = 1000*(4+min_xferp)/(4*hostdata->sync_clock);
747 hostdata->min_period = NCR_700_MIN_PERIOD;
748 if(min_period > NCR_700_MIN_PERIOD)
749 hostdata->min_period = min_period;
750 }
751
752 STATIC void
753 NCR_700_chip_reset(struct Scsi_Host *host)
754 {
755 struct NCR_700_Host_Parameters *hostdata =
756 (struct NCR_700_Host_Parameters *)host->hostdata[0];
757 if(hostdata->chip710) {
758 NCR_700_writeb(SOFTWARE_RESET_710, host, ISTAT_REG);
759 udelay(100);
760
761 NCR_700_writeb(0, host, ISTAT_REG);
762 } else {
763 NCR_700_writeb(SOFTWARE_RESET, host, DCNTL_REG);
764 udelay(100);
765
766 NCR_700_writeb(0, host, DCNTL_REG);
767 }
768
769 mdelay(1000);
770
771 NCR_700_chip_setup(host);
772 }
773
774 /* The heart of the message processing engine is that the instruction
775 * immediately after the INT is the normal case (and so must be CLEAR
776 * ACK). If we want to do something else, we call that routine in
777 * scripts and set temp to be the normal case + 8 (skipping the CLEAR
778 * ACK) so that the routine returns correctly to resume its activity
779 * */
780 STATIC __u32
781 process_extended_message(struct Scsi_Host *host,
782 struct NCR_700_Host_Parameters *hostdata,
783 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
784 {
785 __u32 resume_offset = dsp, temp = dsp + 8;
786 __u8 pun = 0xff, lun = 0xff;
787
788 if(SCp != NULL) {
789 pun = SCp->device->id;
790 lun = SCp->device->lun;
791 }
792
793 switch(hostdata->msgin[2]) {
794 case A_SDTR_MSG:
795 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
796 struct scsi_target *starget = SCp->device->sdev_target;
797 __u8 period = hostdata->msgin[3];
798 __u8 offset = hostdata->msgin[4];
799
800 if(offset == 0 || period == 0) {
801 offset = 0;
802 period = 0;
803 }
804
805 spi_offset(starget) = offset;
806 spi_period(starget) = period;
807
808 if(NCR_700_is_flag_set(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION)) {
809 spi_display_xfer_agreement(starget);
810 NCR_700_clear_flag(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION);
811 }
812
813 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
814 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
815
816 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
817 host, SXFER_REG);
818
819 } else {
820 /* SDTR message out of the blue, reject it */
821 shost_printk(KERN_WARNING, host,
822 "Unexpected SDTR msg\n");
823 hostdata->msgout[0] = A_REJECT_MSG;
824 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
825 script_patch_16(hostdata->script, MessageCount, 1);
826 /* SendMsgOut returns, so set up the return
827 * address */
828 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
829 }
830 break;
831
832 case A_WDTR_MSG:
833 printk(KERN_INFO "scsi%d: (%d:%d), Unsolicited WDTR after CMD, Rejecting\n",
834 host->host_no, pun, lun);
835 hostdata->msgout[0] = A_REJECT_MSG;
836 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
837 script_patch_16(hostdata->script, MessageCount, 1);
838 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
839
840 break;
841
842 default:
843 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
844 host->host_no, pun, lun,
845 NCR_700_phase[(dsps & 0xf00) >> 8]);
846 spi_print_msg(hostdata->msgin);
847 printk("\n");
848 /* just reject it */
849 hostdata->msgout[0] = A_REJECT_MSG;
850 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
851 script_patch_16(hostdata->script, MessageCount, 1);
852 /* SendMsgOut returns, so set up the return
853 * address */
854 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
855 }
856 NCR_700_writel(temp, host, TEMP_REG);
857 return resume_offset;
858 }
859
860 STATIC __u32
861 process_message(struct Scsi_Host *host, struct NCR_700_Host_Parameters *hostdata,
862 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
863 {
864 /* work out where to return to */
865 __u32 temp = dsp + 8, resume_offset = dsp;
866 __u8 pun = 0xff, lun = 0xff;
867
868 if(SCp != NULL) {
869 pun = SCp->device->id;
870 lun = SCp->device->lun;
871 }
872
873 #ifdef NCR_700_DEBUG
874 printk("scsi%d (%d:%d): message %s: ", host->host_no, pun, lun,
875 NCR_700_phase[(dsps & 0xf00) >> 8]);
876 spi_print_msg(hostdata->msgin);
877 printk("\n");
878 #endif
879
880 switch(hostdata->msgin[0]) {
881
882 case A_EXTENDED_MSG:
883 resume_offset = process_extended_message(host, hostdata, SCp,
884 dsp, dsps);
885 break;
886
887 case A_REJECT_MSG:
888 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
889 /* Rejected our sync negotiation attempt */
890 spi_period(SCp->device->sdev_target) =
891 spi_offset(SCp->device->sdev_target) = 0;
892 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
893 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
894 } else if(SCp != NULL && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION) {
895 /* rejected our first simple tag message */
896 scmd_printk(KERN_WARNING, SCp,
897 "Rejected first tag queue attempt, turning off tag queueing\n");
898 /* we're done negotiating */
899 NCR_700_set_tag_neg_state(SCp->device, NCR_700_FINISHED_TAG_NEGOTIATION);
900 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
901 SCp->device->tagged_supported = 0;
902 scsi_deactivate_tcq(SCp->device, host->cmd_per_lun);
903 } else {
904 shost_printk(KERN_WARNING, host,
905 "(%d:%d) Unexpected REJECT Message %s\n",
906 pun, lun,
907 NCR_700_phase[(dsps & 0xf00) >> 8]);
908 /* however, just ignore it */
909 }
910 break;
911
912 case A_PARITY_ERROR_MSG:
913 printk(KERN_ERR "scsi%d (%d:%d) Parity Error!\n", host->host_no,
914 pun, lun);
915 NCR_700_internal_bus_reset(host);
916 break;
917 case A_SIMPLE_TAG_MSG:
918 printk(KERN_INFO "scsi%d (%d:%d) SIMPLE TAG %d %s\n", host->host_no,
919 pun, lun, hostdata->msgin[1],
920 NCR_700_phase[(dsps & 0xf00) >> 8]);
921 /* just ignore it */
922 break;
923 default:
924 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
925 host->host_no, pun, lun,
926 NCR_700_phase[(dsps & 0xf00) >> 8]);
927
928 spi_print_msg(hostdata->msgin);
929 printk("\n");
930 /* just reject it */
931 hostdata->msgout[0] = A_REJECT_MSG;
932 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
933 script_patch_16(hostdata->script, MessageCount, 1);
934 /* SendMsgOut returns, so set up the return
935 * address */
936 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
937
938 break;
939 }
940 NCR_700_writel(temp, host, TEMP_REG);
941 /* set us up to receive another message */
942 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
943 return resume_offset;
944 }
945
946 STATIC __u32
947 process_script_interrupt(__u32 dsps, __u32 dsp, struct scsi_cmnd *SCp,
948 struct Scsi_Host *host,
949 struct NCR_700_Host_Parameters *hostdata)
950 {
951 __u32 resume_offset = 0;
952 __u8 pun = 0xff, lun=0xff;
953
954 if(SCp != NULL) {
955 pun = SCp->device->id;
956 lun = SCp->device->lun;
957 }
958
959 if(dsps == A_GOOD_STATUS_AFTER_STATUS) {
960 DEBUG((" COMMAND COMPLETE, status=%02x\n",
961 hostdata->status[0]));
962 /* OK, if TCQ still under negotiation, we now know it works */
963 if (NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION)
964 NCR_700_set_tag_neg_state(SCp->device,
965 NCR_700_FINISHED_TAG_NEGOTIATION);
966
967 /* check for contingent allegiance contitions */
968 if(status_byte(hostdata->status[0]) == CHECK_CONDITION ||
969 status_byte(hostdata->status[0]) == COMMAND_TERMINATED) {
970 struct NCR_700_command_slot *slot =
971 (struct NCR_700_command_slot *)SCp->host_scribble;
972 if(SCp->cmnd[0] == REQUEST_SENSE) {
973 /* OOPS: bad device, returning another
974 * contingent allegiance condition */
975 scmd_printk(KERN_ERR, SCp,
976 "broken device is looping in contingent allegiance: ignoring\n");
977 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
978 } else {
979 struct NCR_700_sense *sense = SCp->device->hostdata;
980 #ifdef NCR_DEBUG
981 scsi_print_command(SCp);
982 printk(" cmd %p has status %d, requesting sense\n",
983 SCp, hostdata->status[0]);
984 #endif
985 /* we can destroy the command here
986 * because the contingent allegiance
987 * condition will cause a retry which
988 * will re-copy the command from the
989 * saved data_cmnd. We also unmap any
990 * data associated with the command
991 * here */
992 NCR_700_unmap(hostdata, SCp, slot);
993 dma_unmap_single(hostdata->dev, slot->pCmd,
994 sizeof(SCp->cmnd),
995 DMA_TO_DEVICE);
996
997 sense->cmnd[0] = REQUEST_SENSE;
998 sense->cmnd[1] = (SCp->device->lun & 0x7) << 5;
999 sense->cmnd[2] = 0;
1000 sense->cmnd[3] = 0;
1001 sense->cmnd[4] = sizeof(SCp->sense_buffer);
1002 sense->cmnd[5] = 0;
1003 /* Here's a quiet hack: the
1004 * REQUEST_SENSE command is six bytes,
1005 * so store a flag indicating that
1006 * this was an internal sense request
1007 * and the original status at the end
1008 * of the command */
1009 sense->cmnd[6] = NCR_700_INTERNAL_SENSE_MAGIC;
1010 sense->cmnd[7] = hostdata->status[0];
1011 slot->pCmd = dma_map_single(hostdata->dev, sense->cmnd, sizeof(sense->cmnd), DMA_TO_DEVICE);
1012 slot->dma_handle = dma_map_single(hostdata->dev, SCp->sense_buffer, sizeof(SCp->sense_buffer), DMA_FROM_DEVICE);
1013 slot->SG[0].ins = bS_to_host(SCRIPT_MOVE_DATA_IN | sizeof(SCp->sense_buffer));
1014 slot->SG[0].pAddr = bS_to_host(slot->dma_handle);
1015 slot->SG[1].ins = bS_to_host(SCRIPT_RETURN);
1016 slot->SG[1].pAddr = 0;
1017 slot->resume_offset = hostdata->pScript;
1018 dma_cache_sync(slot->SG, sizeof(slot->SG[0])*2, DMA_TO_DEVICE);
1019 dma_cache_sync(SCp->sense_buffer, sizeof(SCp->sense_buffer), DMA_FROM_DEVICE);
1020
1021 /* queue the command for reissue */
1022 slot->state = NCR_700_SLOT_QUEUED;
1023 slot->flags = NCR_700_FLAG_AUTOSENSE;
1024 hostdata->state = NCR_700_HOST_FREE;
1025 hostdata->cmd = NULL;
1026 }
1027 } else {
1028 // Currently rely on the mid layer evaluation
1029 // of the tag queuing capability
1030 //
1031 //if(status_byte(hostdata->status[0]) == GOOD &&
1032 // SCp->cmnd[0] == INQUIRY && SCp->use_sg == 0) {
1033 // /* Piggy back the tag queueing support
1034 // * on this command */
1035 // dma_sync_single_for_cpu(hostdata->dev,
1036 // slot->dma_handle,
1037 // SCp->request_bufflen,
1038 // DMA_FROM_DEVICE);
1039 // if(((char *)SCp->request_buffer)[7] & 0x02) {
1040 // scmd_printk(KERN_INFO, SCp,
1041 // "Enabling Tag Command Queuing\n");
1042 // hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1043 // NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
1044 // } else {
1045 // NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
1046 // hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1047 // }
1048 //}
1049 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
1050 }
1051 } else if((dsps & 0xfffff0f0) == A_UNEXPECTED_PHASE) {
1052 __u8 i = (dsps & 0xf00) >> 8;
1053
1054 scmd_printk(KERN_ERR, SCp, "UNEXPECTED PHASE %s (%s)\n",
1055 NCR_700_phase[i],
1056 sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
1057 scmd_printk(KERN_ERR, SCp, " len = %d, cmd =",
1058 SCp->cmd_len);
1059 scsi_print_command(SCp);
1060
1061 NCR_700_internal_bus_reset(host);
1062 } else if((dsps & 0xfffff000) == A_FATAL) {
1063 int i = (dsps & 0xfff);
1064
1065 printk(KERN_ERR "scsi%d: (%d:%d) FATAL ERROR: %s\n",
1066 host->host_no, pun, lun, NCR_700_fatal_messages[i]);
1067 if(dsps == A_FATAL_ILLEGAL_MSG_LENGTH) {
1068 printk(KERN_ERR " msg begins %02x %02x\n",
1069 hostdata->msgin[0], hostdata->msgin[1]);
1070 }
1071 NCR_700_internal_bus_reset(host);
1072 } else if((dsps & 0xfffff0f0) == A_DISCONNECT) {
1073 #ifdef NCR_700_DEBUG
1074 __u8 i = (dsps & 0xf00) >> 8;
1075
1076 printk("scsi%d: (%d:%d), DISCONNECTED (%d) %s\n",
1077 host->host_no, pun, lun,
1078 i, NCR_700_phase[i]);
1079 #endif
1080 save_for_reselection(hostdata, SCp, dsp);
1081
1082 } else if(dsps == A_RESELECTION_IDENTIFIED) {
1083 __u8 lun;
1084 struct NCR_700_command_slot *slot;
1085 __u8 reselection_id = hostdata->reselection_id;
1086 struct scsi_device *SDp;
1087
1088 lun = hostdata->msgin[0] & 0x1f;
1089
1090 hostdata->reselection_id = 0xff;
1091 DEBUG(("scsi%d: (%d:%d) RESELECTED!\n",
1092 host->host_no, reselection_id, lun));
1093 /* clear the reselection indicator */
1094 SDp = __scsi_device_lookup(host, 0, reselection_id, lun);
1095 if(unlikely(SDp == NULL)) {
1096 printk(KERN_ERR "scsi%d: (%d:%d) HAS NO device\n",
1097 host->host_no, reselection_id, lun);
1098 BUG();
1099 }
1100 if(hostdata->msgin[1] == A_SIMPLE_TAG_MSG) {
1101 struct scsi_cmnd *SCp = scsi_find_tag(SDp, hostdata->msgin[2]);
1102 if(unlikely(SCp == NULL)) {
1103 printk(KERN_ERR "scsi%d: (%d:%d) no saved request for tag %d\n",
1104 host->host_no, reselection_id, lun, hostdata->msgin[2]);
1105 BUG();
1106 }
1107
1108 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1109 DDEBUG(KERN_DEBUG, SDp,
1110 "reselection is tag %d, slot %p(%d)\n",
1111 hostdata->msgin[2], slot, slot->tag);
1112 } else {
1113 struct scsi_cmnd *SCp = scsi_find_tag(SDp, SCSI_NO_TAG);
1114 if(unlikely(SCp == NULL)) {
1115 sdev_printk(KERN_ERR, SDp,
1116 "no saved request for untagged cmd\n");
1117 BUG();
1118 }
1119 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1120 }
1121
1122 if(slot == NULL) {
1123 printk(KERN_ERR "scsi%d: (%d:%d) RESELECTED but no saved command (MSG = %02x %02x %02x)!!\n",
1124 host->host_no, reselection_id, lun,
1125 hostdata->msgin[0], hostdata->msgin[1],
1126 hostdata->msgin[2]);
1127 } else {
1128 if(hostdata->state != NCR_700_HOST_BUSY)
1129 printk(KERN_ERR "scsi%d: FATAL, host not busy during valid reselection!\n",
1130 host->host_no);
1131 resume_offset = slot->resume_offset;
1132 hostdata->cmd = slot->cmnd;
1133
1134 /* re-patch for this command */
1135 script_patch_32_abs(hostdata->script, CommandAddress,
1136 slot->pCmd);
1137 script_patch_16(hostdata->script,
1138 CommandCount, slot->cmnd->cmd_len);
1139 script_patch_32_abs(hostdata->script, SGScriptStartAddress,
1140 to32bit(&slot->pSG[0].ins));
1141
1142 /* Note: setting SXFER only works if we're
1143 * still in the MESSAGE phase, so it is vital
1144 * that ACK is still asserted when we process
1145 * the reselection message. The resume offset
1146 * should therefore always clear ACK */
1147 NCR_700_writeb(NCR_700_get_SXFER(hostdata->cmd->device),
1148 host, SXFER_REG);
1149 dma_cache_sync(hostdata->msgin,
1150 MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
1151 dma_cache_sync(hostdata->msgout,
1152 MSG_ARRAY_SIZE, DMA_TO_DEVICE);
1153 /* I'm just being paranoid here, the command should
1154 * already have been flushed from the cache */
1155 dma_cache_sync(slot->cmnd->cmnd,
1156 slot->cmnd->cmd_len, DMA_TO_DEVICE);
1157
1158
1159
1160 }
1161 } else if(dsps == A_RESELECTED_DURING_SELECTION) {
1162
1163 /* This section is full of debugging code because I've
1164 * never managed to reach it. I think what happens is
1165 * that, because the 700 runs with selection
1166 * interrupts enabled the whole time that we take a
1167 * selection interrupt before we manage to get to the
1168 * reselected script interrupt */
1169
1170 __u8 reselection_id = NCR_700_readb(host, SFBR_REG);
1171 struct NCR_700_command_slot *slot;
1172
1173 /* Take out our own ID */
1174 reselection_id &= ~(1<<host->this_id);
1175
1176 /* I've never seen this happen, so keep this as a printk rather
1177 * than a debug */
1178 printk(KERN_INFO "scsi%d: (%d:%d) RESELECTION DURING SELECTION, dsp=%08x[%04x] state=%d, count=%d\n",
1179 host->host_no, reselection_id, lun, dsp, dsp - hostdata->pScript, hostdata->state, hostdata->command_slot_count);
1180
1181 {
1182 /* FIXME: DEBUGGING CODE */
1183 __u32 SG = (__u32)bS_to_cpu(hostdata->script[A_SGScriptStartAddress_used[0]]);
1184 int i;
1185
1186 for(i=0; i< NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1187 if(SG >= to32bit(&hostdata->slots[i].pSG[0])
1188 && SG <= to32bit(&hostdata->slots[i].pSG[NCR_700_SG_SEGMENTS]))
1189 break;
1190 }
1191 printk(KERN_INFO "IDENTIFIED SG segment as being %08x in slot %p, cmd %p, slot->resume_offset=%08x\n", SG, &hostdata->slots[i], hostdata->slots[i].cmnd, hostdata->slots[i].resume_offset);
1192 SCp = hostdata->slots[i].cmnd;
1193 }
1194
1195 if(SCp != NULL) {
1196 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1197 /* change slot from busy to queued to redo command */
1198 slot->state = NCR_700_SLOT_QUEUED;
1199 }
1200 hostdata->cmd = NULL;
1201
1202 if(reselection_id == 0) {
1203 if(hostdata->reselection_id == 0xff) {
1204 printk(KERN_ERR "scsi%d: Invalid reselection during selection!!\n", host->host_no);
1205 return 0;
1206 } else {
1207 printk(KERN_ERR "scsi%d: script reselected and we took a selection interrupt\n",
1208 host->host_no);
1209 reselection_id = hostdata->reselection_id;
1210 }
1211 } else {
1212
1213 /* convert to real ID */
1214 reselection_id = bitmap_to_number(reselection_id);
1215 }
1216 hostdata->reselection_id = reselection_id;
1217 /* just in case we have a stale simple tag message, clear it */
1218 hostdata->msgin[1] = 0;
1219 dma_cache_sync(hostdata->msgin,
1220 MSG_ARRAY_SIZE, DMA_BIDIRECTIONAL);
1221 if(hostdata->tag_negotiated & (1<<reselection_id)) {
1222 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1223 } else {
1224 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1225 }
1226 } else if(dsps == A_COMPLETED_SELECTION_AS_TARGET) {
1227 /* we've just disconnected from the bus, do nothing since
1228 * a return here will re-run the queued command slot
1229 * that may have been interrupted by the initial selection */
1230 DEBUG((" SELECTION COMPLETED\n"));
1231 } else if((dsps & 0xfffff0f0) == A_MSG_IN) {
1232 resume_offset = process_message(host, hostdata, SCp,
1233 dsp, dsps);
1234 } else if((dsps & 0xfffff000) == 0) {
1235 __u8 i = (dsps & 0xf0) >> 4, j = (dsps & 0xf00) >> 8;
1236 printk(KERN_ERR "scsi%d: (%d:%d), unhandled script condition %s %s at %04x\n",
1237 host->host_no, pun, lun, NCR_700_condition[i],
1238 NCR_700_phase[j], dsp - hostdata->pScript);
1239 if(SCp != NULL) {
1240 scsi_print_command(SCp);
1241
1242 if(SCp->use_sg) {
1243 for(i = 0; i < SCp->use_sg + 1; i++) {
1244 printk(KERN_INFO " SG[%d].length = %d, move_insn=%08x, addr %08x\n", i, ((struct scatterlist *)SCp->request_buffer)[i].length, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].ins, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].pAddr);
1245 }
1246 }
1247 }
1248 NCR_700_internal_bus_reset(host);
1249 } else if((dsps & 0xfffff000) == A_DEBUG_INTERRUPT) {
1250 printk(KERN_NOTICE "scsi%d (%d:%d) DEBUG INTERRUPT %d AT %08x[%04x], continuing\n",
1251 host->host_no, pun, lun, dsps & 0xfff, dsp, dsp - hostdata->pScript);
1252 resume_offset = dsp;
1253 } else {
1254 printk(KERN_ERR "scsi%d: (%d:%d), unidentified script interrupt 0x%x at %04x\n",
1255 host->host_no, pun, lun, dsps, dsp - hostdata->pScript);
1256 NCR_700_internal_bus_reset(host);
1257 }
1258 return resume_offset;
1259 }
1260
1261 /* We run the 53c700 with selection interrupts always enabled. This
1262 * means that the chip may be selected as soon as the bus frees. On a
1263 * busy bus, this can be before the scripts engine finishes its
1264 * processing. Therefore, part of the selection processing has to be
1265 * to find out what the scripts engine is doing and complete the
1266 * function if necessary (i.e. process the pending disconnect or save
1267 * the interrupted initial selection */
1268 STATIC inline __u32
1269 process_selection(struct Scsi_Host *host, __u32 dsp)
1270 {
1271 __u8 id = 0; /* Squash compiler warning */
1272 int count = 0;
1273 __u32 resume_offset = 0;
1274 struct NCR_700_Host_Parameters *hostdata =
1275 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1276 struct scsi_cmnd *SCp = hostdata->cmd;
1277 __u8 sbcl;
1278
1279 for(count = 0; count < 5; count++) {
1280 id = NCR_700_readb(host, hostdata->chip710 ?
1281 CTEST9_REG : SFBR_REG);
1282
1283 /* Take out our own ID */
1284 id &= ~(1<<host->this_id);
1285 if(id != 0)
1286 break;
1287 udelay(5);
1288 }
1289 sbcl = NCR_700_readb(host, SBCL_REG);
1290 if((sbcl & SBCL_IO) == 0) {
1291 /* mark as having been selected rather than reselected */
1292 id = 0xff;
1293 } else {
1294 /* convert to real ID */
1295 hostdata->reselection_id = id = bitmap_to_number(id);
1296 DEBUG(("scsi%d: Reselected by %d\n",
1297 host->host_no, id));
1298 }
1299 if(hostdata->state == NCR_700_HOST_BUSY && SCp != NULL) {
1300 struct NCR_700_command_slot *slot =
1301 (struct NCR_700_command_slot *)SCp->host_scribble;
1302 DEBUG((" ID %d WARNING: RESELECTION OF BUSY HOST, saving cmd %p, slot %p, addr %x [%04x], resume %x!\n", id, hostdata->cmd, slot, dsp, dsp - hostdata->pScript, resume_offset));
1303
1304 switch(dsp - hostdata->pScript) {
1305 case Ent_Disconnect1:
1306 case Ent_Disconnect2:
1307 save_for_reselection(hostdata, SCp, Ent_Disconnect2 + hostdata->pScript);
1308 break;
1309 case Ent_Disconnect3:
1310 case Ent_Disconnect4:
1311 save_for_reselection(hostdata, SCp, Ent_Disconnect4 + hostdata->pScript);
1312 break;
1313 case Ent_Disconnect5:
1314 case Ent_Disconnect6:
1315 save_for_reselection(hostdata, SCp, Ent_Disconnect6 + hostdata->pScript);
1316 break;
1317 case Ent_Disconnect7:
1318 case Ent_Disconnect8:
1319 save_for_reselection(hostdata, SCp, Ent_Disconnect8 + hostdata->pScript);
1320 break;
1321 case Ent_Finish1:
1322 case Ent_Finish2:
1323 process_script_interrupt(A_GOOD_STATUS_AFTER_STATUS, dsp, SCp, host, hostdata);
1324 break;
1325
1326 default:
1327 slot->state = NCR_700_SLOT_QUEUED;
1328 break;
1329 }
1330 }
1331 hostdata->state = NCR_700_HOST_BUSY;
1332 hostdata->cmd = NULL;
1333 /* clear any stale simple tag message */
1334 hostdata->msgin[1] = 0;
1335 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE,
1336 DMA_BIDIRECTIONAL);
1337
1338 if(id == 0xff) {
1339 /* Selected as target, Ignore */
1340 resume_offset = hostdata->pScript + Ent_SelectedAsTarget;
1341 } else if(hostdata->tag_negotiated & (1<<id)) {
1342 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1343 } else {
1344 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1345 }
1346 return resume_offset;
1347 }
1348
1349 static inline void
1350 NCR_700_clear_fifo(struct Scsi_Host *host) {
1351 const struct NCR_700_Host_Parameters *hostdata
1352 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1353 if(hostdata->chip710) {
1354 NCR_700_writeb(CLR_FIFO_710, host, CTEST8_REG);
1355 } else {
1356 NCR_700_writeb(CLR_FIFO, host, DFIFO_REG);
1357 }
1358 }
1359
1360 static inline void
1361 NCR_700_flush_fifo(struct Scsi_Host *host) {
1362 const struct NCR_700_Host_Parameters *hostdata
1363 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1364 if(hostdata->chip710) {
1365 NCR_700_writeb(FLUSH_DMA_FIFO_710, host, CTEST8_REG);
1366 udelay(10);
1367 NCR_700_writeb(0, host, CTEST8_REG);
1368 } else {
1369 NCR_700_writeb(FLUSH_DMA_FIFO, host, DFIFO_REG);
1370 udelay(10);
1371 NCR_700_writeb(0, host, DFIFO_REG);
1372 }
1373 }
1374
1375
1376 /* The queue lock with interrupts disabled must be held on entry to
1377 * this function */
1378 STATIC int
1379 NCR_700_start_command(struct scsi_cmnd *SCp)
1380 {
1381 struct NCR_700_command_slot *slot =
1382 (struct NCR_700_command_slot *)SCp->host_scribble;
1383 struct NCR_700_Host_Parameters *hostdata =
1384 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1385 __u16 count = 1; /* for IDENTIFY message */
1386
1387 if(hostdata->state != NCR_700_HOST_FREE) {
1388 /* keep this inside the lock to close the race window where
1389 * the running command finishes on another CPU while we don't
1390 * change the state to queued on this one */
1391 slot->state = NCR_700_SLOT_QUEUED;
1392
1393 DEBUG(("scsi%d: host busy, queueing command %p, slot %p\n",
1394 SCp->device->host->host_no, slot->cmnd, slot));
1395 return 0;
1396 }
1397 hostdata->state = NCR_700_HOST_BUSY;
1398 hostdata->cmd = SCp;
1399 slot->state = NCR_700_SLOT_BUSY;
1400 /* keep interrupts disabled until we have the command correctly
1401 * set up so we cannot take a selection interrupt */
1402
1403 hostdata->msgout[0] = NCR_700_identify((SCp->cmnd[0] != REQUEST_SENSE &&
1404 slot->flags != NCR_700_FLAG_AUTOSENSE),
1405 SCp->device->lun);
1406 /* for INQUIRY or REQUEST_SENSE commands, we cannot be sure
1407 * if the negotiated transfer parameters still hold, so
1408 * always renegotiate them */
1409 if(SCp->cmnd[0] == INQUIRY || SCp->cmnd[0] == REQUEST_SENSE ||
1410 slot->flags == NCR_700_FLAG_AUTOSENSE) {
1411 NCR_700_clear_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
1412 }
1413
1414 /* REQUEST_SENSE is asking for contingent I_T_L(_Q) status.
1415 * If a contingent allegiance condition exists, the device
1416 * will refuse all tags, so send the request sense as untagged
1417 * */
1418 if((hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1419 && (slot->tag != SCSI_NO_TAG && SCp->cmnd[0] != REQUEST_SENSE &&
1420 slot->flags != NCR_700_FLAG_AUTOSENSE)) {
1421 count += scsi_populate_tag_msg(SCp, &hostdata->msgout[count]);
1422 }
1423
1424 if(hostdata->fast &&
1425 NCR_700_is_flag_clear(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC)) {
1426 count += spi_populate_sync_msg(&hostdata->msgout[count],
1427 spi_period(SCp->device->sdev_target),
1428 spi_offset(SCp->device->sdev_target));
1429 NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
1430 }
1431
1432 script_patch_16(hostdata->script, MessageCount, count);
1433
1434
1435 script_patch_ID(hostdata->script,
1436 Device_ID, 1<<scmd_id(SCp));
1437
1438 script_patch_32_abs(hostdata->script, CommandAddress,
1439 slot->pCmd);
1440 script_patch_16(hostdata->script, CommandCount, SCp->cmd_len);
1441 /* finally plumb the beginning of the SG list into the script
1442 * */
1443 script_patch_32_abs(hostdata->script, SGScriptStartAddress,
1444 to32bit(&slot->pSG[0].ins));
1445 NCR_700_clear_fifo(SCp->device->host);
1446
1447 if(slot->resume_offset == 0)
1448 slot->resume_offset = hostdata->pScript;
1449 /* now perform all the writebacks and invalidates */
1450 dma_cache_sync(hostdata->msgout, count, DMA_TO_DEVICE);
1451 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE,
1452 DMA_FROM_DEVICE);
1453 dma_cache_sync(SCp->cmnd, SCp->cmd_len, DMA_TO_DEVICE);
1454 dma_cache_sync(hostdata->status, 1, DMA_FROM_DEVICE);
1455
1456 /* set the synchronous period/offset */
1457 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
1458 SCp->device->host, SXFER_REG);
1459 NCR_700_writel(slot->temp, SCp->device->host, TEMP_REG);
1460 NCR_700_writel(slot->resume_offset, SCp->device->host, DSP_REG);
1461
1462 return 1;
1463 }
1464
1465 irqreturn_t
1466 NCR_700_intr(int irq, void *dev_id, struct pt_regs *regs)
1467 {
1468 struct Scsi_Host *host = (struct Scsi_Host *)dev_id;
1469 struct NCR_700_Host_Parameters *hostdata =
1470 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1471 __u8 istat;
1472 __u32 resume_offset = 0;
1473 __u8 pun = 0xff, lun = 0xff;
1474 unsigned long flags;
1475 int handled = 0;
1476
1477 /* Use the host lock to serialise acess to the 53c700
1478 * hardware. Note: In future, we may need to take the queue
1479 * lock to enter the done routines. When that happens, we
1480 * need to ensure that for this driver, the host lock and the
1481 * queue lock point to the same thing. */
1482 spin_lock_irqsave(host->host_lock, flags);
1483 if((istat = NCR_700_readb(host, ISTAT_REG))
1484 & (SCSI_INT_PENDING | DMA_INT_PENDING)) {
1485 __u32 dsps;
1486 __u8 sstat0 = 0, dstat = 0;
1487 __u32 dsp;
1488 struct scsi_cmnd *SCp = hostdata->cmd;
1489 enum NCR_700_Host_State state;
1490
1491 handled = 1;
1492 state = hostdata->state;
1493 SCp = hostdata->cmd;
1494
1495 if(istat & SCSI_INT_PENDING) {
1496 udelay(10);
1497
1498 sstat0 = NCR_700_readb(host, SSTAT0_REG);
1499 }
1500
1501 if(istat & DMA_INT_PENDING) {
1502 udelay(10);
1503
1504 dstat = NCR_700_readb(host, DSTAT_REG);
1505 }
1506
1507 dsps = NCR_700_readl(host, DSPS_REG);
1508 dsp = NCR_700_readl(host, DSP_REG);
1509
1510 DEBUG(("scsi%d: istat %02x sstat0 %02x dstat %02x dsp %04x[%08x] dsps 0x%x\n",
1511 host->host_no, istat, sstat0, dstat,
1512 (dsp - (__u32)(hostdata->pScript))/4,
1513 dsp, dsps));
1514
1515 if(SCp != NULL) {
1516 pun = SCp->device->id;
1517 lun = SCp->device->lun;
1518 }
1519
1520 if(sstat0 & SCSI_RESET_DETECTED) {
1521 struct scsi_device *SDp;
1522 int i;
1523
1524 hostdata->state = NCR_700_HOST_BUSY;
1525
1526 printk(KERN_ERR "scsi%d: Bus Reset detected, executing command %p, slot %p, dsp %08x[%04x]\n",
1527 host->host_no, SCp, SCp == NULL ? NULL : SCp->host_scribble, dsp, dsp - hostdata->pScript);
1528
1529 scsi_report_bus_reset(host, 0);
1530
1531 /* clear all the negotiated parameters */
1532 __shost_for_each_device(SDp, host)
1533 SDp->hostdata = NULL;
1534
1535 /* clear all the slots and their pending commands */
1536 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1537 struct scsi_cmnd *SCp;
1538 struct NCR_700_command_slot *slot =
1539 &hostdata->slots[i];
1540
1541 if(slot->state == NCR_700_SLOT_FREE)
1542 continue;
1543
1544 SCp = slot->cmnd;
1545 printk(KERN_ERR " failing command because of reset, slot %p, cmnd %p\n",
1546 slot, SCp);
1547 free_slot(slot, hostdata);
1548 SCp->host_scribble = NULL;
1549 NCR_700_set_depth(SCp->device, 0);
1550 /* NOTE: deadlock potential here: we
1551 * rely on mid-layer guarantees that
1552 * scsi_done won't try to issue the
1553 * command again otherwise we'll
1554 * deadlock on the
1555 * hostdata->state_lock */
1556 SCp->result = DID_RESET << 16;
1557 SCp->scsi_done(SCp);
1558 }
1559 mdelay(25);
1560 NCR_700_chip_setup(host);
1561
1562 hostdata->state = NCR_700_HOST_FREE;
1563 hostdata->cmd = NULL;
1564 /* signal back if this was an eh induced reset */
1565 if(hostdata->eh_complete != NULL)
1566 complete(hostdata->eh_complete);
1567 goto out_unlock;
1568 } else if(sstat0 & SELECTION_TIMEOUT) {
1569 DEBUG(("scsi%d: (%d:%d) selection timeout\n",
1570 host->host_no, pun, lun));
1571 NCR_700_scsi_done(hostdata, SCp, DID_NO_CONNECT<<16);
1572 } else if(sstat0 & PHASE_MISMATCH) {
1573 struct NCR_700_command_slot *slot = (SCp == NULL) ? NULL :
1574 (struct NCR_700_command_slot *)SCp->host_scribble;
1575
1576 if(dsp == Ent_SendMessage + 8 + hostdata->pScript) {
1577 /* It wants to reply to some part of
1578 * our message */
1579 #ifdef NCR_700_DEBUG
1580 __u32 temp = NCR_700_readl(host, TEMP_REG);
1581 int count = (hostdata->script[Ent_SendMessage/4] & 0xffffff) - ((NCR_700_readl(host, DBC_REG) & 0xffffff) + NCR_700_data_residual(host));
1582 printk("scsi%d (%d:%d) PHASE MISMATCH IN SEND MESSAGE %d remain, return %p[%04x], phase %s\n", host->host_no, pun, lun, count, (void *)temp, temp - hostdata->pScript, sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
1583 #endif
1584 resume_offset = hostdata->pScript + Ent_SendMessagePhaseMismatch;
1585 } else if(dsp >= to32bit(&slot->pSG[0].ins) &&
1586 dsp <= to32bit(&slot->pSG[NCR_700_SG_SEGMENTS].ins)) {
1587 int data_transfer = NCR_700_readl(host, DBC_REG) & 0xffffff;
1588 int SGcount = (dsp - to32bit(&slot->pSG[0].ins))/sizeof(struct NCR_700_SG_List);
1589 int residual = NCR_700_data_residual(host);
1590 int i;
1591 #ifdef NCR_700_DEBUG
1592 __u32 naddr = NCR_700_readl(host, DNAD_REG);
1593
1594 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x\n",
1595 host->host_no, pun, lun,
1596 SGcount, data_transfer);
1597 scsi_print_command(SCp);
1598 if(residual) {
1599 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x, residual %d\n",
1600 host->host_no, pun, lun,
1601 SGcount, data_transfer, residual);
1602 }
1603 #endif
1604 data_transfer += residual;
1605
1606 if(data_transfer != 0) {
1607 int count;
1608 __u32 pAddr;
1609
1610 SGcount--;
1611
1612 count = (bS_to_cpu(slot->SG[SGcount].ins) & 0x00ffffff);
1613 DEBUG(("DATA TRANSFER MISMATCH, count = %d, transferred %d\n", count, count-data_transfer));
1614 slot->SG[SGcount].ins &= bS_to_host(0xff000000);
1615 slot->SG[SGcount].ins |= bS_to_host(data_transfer);
1616 pAddr = bS_to_cpu(slot->SG[SGcount].pAddr);
1617 pAddr += (count - data_transfer);
1618 #ifdef NCR_700_DEBUG
1619 if(pAddr != naddr) {
1620 printk("scsi%d (%d:%d) transfer mismatch pAddr=%lx, naddr=%lx, data_transfer=%d, residual=%d\n", host->host_no, pun, lun, (unsigned long)pAddr, (unsigned long)naddr, data_transfer, residual);
1621 }
1622 #endif
1623 slot->SG[SGcount].pAddr = bS_to_host(pAddr);
1624 }
1625 /* set the executed moves to nops */
1626 for(i=0; i<SGcount; i++) {
1627 slot->SG[i].ins = bS_to_host(SCRIPT_NOP);
1628 slot->SG[i].pAddr = 0;
1629 }
1630 dma_cache_sync(slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1631 /* and pretend we disconnected after
1632 * the command phase */
1633 resume_offset = hostdata->pScript + Ent_MsgInDuringData;
1634 /* make sure all the data is flushed */
1635 NCR_700_flush_fifo(host);
1636 } else {
1637 __u8 sbcl = NCR_700_readb(host, SBCL_REG);
1638 printk(KERN_ERR "scsi%d: (%d:%d) phase mismatch at %04x, phase %s\n",
1639 host->host_no, pun, lun, dsp - hostdata->pScript, sbcl_to_string(sbcl));
1640 NCR_700_internal_bus_reset(host);
1641 }
1642
1643 } else if(sstat0 & SCSI_GROSS_ERROR) {
1644 printk(KERN_ERR "scsi%d: (%d:%d) GROSS ERROR\n",
1645 host->host_no, pun, lun);
1646 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1647 } else if(sstat0 & PARITY_ERROR) {
1648 printk(KERN_ERR "scsi%d: (%d:%d) PARITY ERROR\n",
1649 host->host_no, pun, lun);
1650 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1651 } else if(dstat & SCRIPT_INT_RECEIVED) {
1652 DEBUG(("scsi%d: (%d:%d) ====>SCRIPT INTERRUPT<====\n",
1653 host->host_no, pun, lun));
1654 resume_offset = process_script_interrupt(dsps, dsp, SCp, host, hostdata);
1655 } else if(dstat & (ILGL_INST_DETECTED)) {
1656 printk(KERN_ERR "scsi%d: (%d:%d) Illegal Instruction detected at 0x%08x[0x%x]!!!\n"
1657 " Please email James.Bottomley@HansenPartnership.com with the details\n",
1658 host->host_no, pun, lun,
1659 dsp, dsp - hostdata->pScript);
1660 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1661 } else if(dstat & (WATCH_DOG_INTERRUPT|ABORTED)) {
1662 printk(KERN_ERR "scsi%d: (%d:%d) serious DMA problem, dstat=%02x\n",
1663 host->host_no, pun, lun, dstat);
1664 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1665 }
1666
1667
1668 /* NOTE: selection interrupt processing MUST occur
1669 * after script interrupt processing to correctly cope
1670 * with the case where we process a disconnect and
1671 * then get reselected before we process the
1672 * disconnection */
1673 if(sstat0 & SELECTED) {
1674 /* FIXME: It currently takes at least FOUR
1675 * interrupts to complete a command that
1676 * disconnects: one for the disconnect, one
1677 * for the reselection, one to get the
1678 * reselection data and one to complete the
1679 * command. If we guess the reselected
1680 * command here and prepare it, we only need
1681 * to get a reselection data interrupt if we
1682 * guessed wrongly. Since the interrupt
1683 * overhead is much greater than the command
1684 * setup, this would be an efficient
1685 * optimisation particularly as we probably
1686 * only have one outstanding command on a
1687 * target most of the time */
1688
1689 resume_offset = process_selection(host, dsp);
1690
1691 }
1692
1693 }
1694
1695 if(resume_offset) {
1696 if(hostdata->state != NCR_700_HOST_BUSY) {
1697 printk(KERN_ERR "scsi%d: Driver error: resume at 0x%08x [0x%04x] with non busy host!\n",
1698 host->host_no, resume_offset, resume_offset - hostdata->pScript);
1699 hostdata->state = NCR_700_HOST_BUSY;
1700 }
1701
1702 DEBUG(("Attempting to resume at %x\n", resume_offset));
1703 NCR_700_clear_fifo(host);
1704 NCR_700_writel(resume_offset, host, DSP_REG);
1705 }
1706 /* There is probably a technical no-no about this: If we're a
1707 * shared interrupt and we got this interrupt because the
1708 * other device needs servicing not us, we're still going to
1709 * check our queued commands here---of course, there shouldn't
1710 * be any outstanding.... */
1711 if(hostdata->state == NCR_700_HOST_FREE) {
1712 int i;
1713
1714 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1715 /* fairness: always run the queue from the last
1716 * position we left off */
1717 int j = (i + hostdata->saved_slot_position)
1718 % NCR_700_COMMAND_SLOTS_PER_HOST;
1719
1720 if(hostdata->slots[j].state != NCR_700_SLOT_QUEUED)
1721 continue;
1722 if(NCR_700_start_command(hostdata->slots[j].cmnd)) {
1723 DEBUG(("scsi%d: Issuing saved command slot %p, cmd %p\t\n",
1724 host->host_no, &hostdata->slots[j],
1725 hostdata->slots[j].cmnd));
1726 hostdata->saved_slot_position = j + 1;
1727 }
1728
1729 break;
1730 }
1731 }
1732 out_unlock:
1733 spin_unlock_irqrestore(host->host_lock, flags);
1734 return IRQ_RETVAL(handled);
1735 }
1736
1737 STATIC int
1738 NCR_700_queuecommand(struct scsi_cmnd *SCp, void (*done)(struct scsi_cmnd *))
1739 {
1740 struct NCR_700_Host_Parameters *hostdata =
1741 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1742 __u32 move_ins;
1743 enum dma_data_direction direction;
1744 struct NCR_700_command_slot *slot;
1745
1746 if(hostdata->command_slot_count >= NCR_700_COMMAND_SLOTS_PER_HOST) {
1747 /* We're over our allocation, this should never happen
1748 * since we report the max allocation to the mid layer */
1749 printk(KERN_WARNING "scsi%d: Command depth has gone over queue depth\n", SCp->device->host->host_no);
1750 return 1;
1751 }
1752 /* check for untagged commands. We cannot have any outstanding
1753 * commands if we accept them. Commands could be untagged because:
1754 *
1755 * - The tag negotiated bitmap is clear
1756 * - The blk layer sent and untagged command
1757 */
1758 if(NCR_700_get_depth(SCp->device) != 0
1759 && (!(hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1760 || !blk_rq_tagged(SCp->request))) {
1761 CDEBUG(KERN_ERR, SCp, "has non zero depth %d\n",
1762 NCR_700_get_depth(SCp->device));
1763 return SCSI_MLQUEUE_DEVICE_BUSY;
1764 }
1765 if(NCR_700_get_depth(SCp->device) >= SCp->device->queue_depth) {
1766 CDEBUG(KERN_ERR, SCp, "has max tag depth %d\n",
1767 NCR_700_get_depth(SCp->device));
1768 return SCSI_MLQUEUE_DEVICE_BUSY;
1769 }
1770 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) + 1);
1771
1772 /* begin the command here */
1773 /* no need to check for NULL, test for command_slot_count above
1774 * ensures a slot is free */
1775 slot = find_empty_slot(hostdata);
1776
1777 slot->cmnd = SCp;
1778
1779 SCp->scsi_done = done;
1780 SCp->host_scribble = (unsigned char *)slot;
1781 SCp->SCp.ptr = NULL;
1782 SCp->SCp.buffer = NULL;
1783
1784 #ifdef NCR_700_DEBUG
1785 printk("53c700: scsi%d, command ", SCp->device->host->host_no);
1786 scsi_print_command(SCp);
1787 #endif
1788 if(blk_rq_tagged(SCp->request)
1789 && (hostdata->tag_negotiated &(1<<scmd_id(SCp))) == 0
1790 && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_START_TAG_NEGOTIATION) {
1791 scmd_printk(KERN_ERR, SCp, "Enabling Tag Command Queuing\n");
1792 hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1793 NCR_700_set_tag_neg_state(SCp->device, NCR_700_DURING_TAG_NEGOTIATION);
1794 }
1795
1796 /* here we may have to process an untagged command. The gate
1797 * above ensures that this will be the only one outstanding,
1798 * so clear the tag negotiated bit.
1799 *
1800 * FIXME: This will royally screw up on multiple LUN devices
1801 * */
1802 if(!blk_rq_tagged(SCp->request)
1803 && (hostdata->tag_negotiated &(1<<scmd_id(SCp)))) {
1804 scmd_printk(KERN_INFO, SCp, "Disabling Tag Command Queuing\n");
1805 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1806 }
1807
1808 if((hostdata->tag_negotiated &(1<<scmd_id(SCp)))
1809 && scsi_get_tag_type(SCp->device)) {
1810 slot->tag = SCp->request->tag;
1811 CDEBUG(KERN_DEBUG, SCp, "sending out tag %d, slot %p\n",
1812 slot->tag, slot);
1813 } else {
1814 slot->tag = SCSI_NO_TAG;
1815 /* must populate current_cmnd for scsi_find_tag to work */
1816 SCp->device->current_cmnd = SCp;
1817 }
1818 /* sanity check: some of the commands generated by the mid-layer
1819 * have an eccentric idea of their sc_data_direction */
1820 if(!SCp->use_sg && !SCp->request_bufflen
1821 && SCp->sc_data_direction != DMA_NONE) {
1822 #ifdef NCR_700_DEBUG
1823 printk("53c700: Command");
1824 scsi_print_command(SCp);
1825 printk("Has wrong data direction %d\n", SCp->sc_data_direction);
1826 #endif
1827 SCp->sc_data_direction = DMA_NONE;
1828 }
1829
1830 switch (SCp->cmnd[0]) {
1831 case REQUEST_SENSE:
1832 /* clear the internal sense magic */
1833 SCp->cmnd[6] = 0;
1834 /* fall through */
1835 default:
1836 /* OK, get it from the command */
1837 switch(SCp->sc_data_direction) {
1838 case DMA_BIDIRECTIONAL:
1839 default:
1840 printk(KERN_ERR "53c700: Unknown command for data direction ");
1841 scsi_print_command(SCp);
1842
1843 move_ins = 0;
1844 break;
1845 case DMA_NONE:
1846 move_ins = 0;
1847 break;
1848 case DMA_FROM_DEVICE:
1849 move_ins = SCRIPT_MOVE_DATA_IN;
1850 break;
1851 case DMA_TO_DEVICE:
1852 move_ins = SCRIPT_MOVE_DATA_OUT;
1853 break;
1854 }
1855 }
1856
1857 /* now build the scatter gather list */
1858 direction = SCp->sc_data_direction;
1859 if(move_ins != 0) {
1860 int i;
1861 int sg_count;
1862 dma_addr_t vPtr = 0;
1863 __u32 count = 0;
1864
1865 if(SCp->use_sg) {
1866 sg_count = dma_map_sg(hostdata->dev,
1867 SCp->request_buffer, SCp->use_sg,
1868 direction);
1869 } else {
1870 vPtr = dma_map_single(hostdata->dev,
1871 SCp->request_buffer,
1872 SCp->request_bufflen,
1873 direction);
1874 count = SCp->request_bufflen;
1875 slot->dma_handle = vPtr;
1876 sg_count = 1;
1877 }
1878
1879
1880 for(i = 0; i < sg_count; i++) {
1881
1882 if(SCp->use_sg) {
1883 struct scatterlist *sg = SCp->request_buffer;
1884
1885 vPtr = sg_dma_address(&sg[i]);
1886 count = sg_dma_len(&sg[i]);
1887 }
1888
1889 slot->SG[i].ins = bS_to_host(move_ins | count);
1890 DEBUG((" scatter block %d: move %d[%08x] from 0x%lx\n",
1891 i, count, slot->SG[i].ins, (unsigned long)vPtr));
1892 slot->SG[i].pAddr = bS_to_host(vPtr);
1893 }
1894 slot->SG[i].ins = bS_to_host(SCRIPT_RETURN);
1895 slot->SG[i].pAddr = 0;
1896 dma_cache_sync(slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1897 DEBUG((" SETTING %08lx to %x\n",
1898 (&slot->pSG[i].ins),
1899 slot->SG[i].ins));
1900 }
1901 slot->resume_offset = 0;
1902 slot->pCmd = dma_map_single(hostdata->dev, SCp->cmnd,
1903 sizeof(SCp->cmnd), DMA_TO_DEVICE);
1904 NCR_700_start_command(SCp);
1905 return 0;
1906 }
1907
1908 STATIC int
1909 NCR_700_abort(struct scsi_cmnd * SCp)
1910 {
1911 struct NCR_700_command_slot *slot;
1912
1913 scmd_printk(KERN_INFO, SCp,
1914 "New error handler wants to abort command\n\t");
1915 scsi_print_command(SCp);
1916
1917 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1918
1919 if(slot == NULL)
1920 /* no outstanding command to abort */
1921 return SUCCESS;
1922 if(SCp->cmnd[0] == TEST_UNIT_READY) {
1923 /* FIXME: This is because of a problem in the new
1924 * error handler. When it is in error recovery, it
1925 * will send a TUR to a device it thinks may still be
1926 * showing a problem. If the TUR isn't responded to,
1927 * it will abort it and mark the device off line.
1928 * Unfortunately, it does no other error recovery, so
1929 * this would leave us with an outstanding command
1930 * occupying a slot. Rather than allow this to
1931 * happen, we issue a bus reset to force all
1932 * outstanding commands to terminate here. */
1933 NCR_700_internal_bus_reset(SCp->device->host);
1934 /* still drop through and return failed */
1935 }
1936 return FAILED;
1937
1938 }
1939
1940 STATIC int
1941 NCR_700_bus_reset(struct scsi_cmnd * SCp)
1942 {
1943 DECLARE_COMPLETION(complete);
1944 struct NCR_700_Host_Parameters *hostdata =
1945 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1946
1947 scmd_printk(KERN_INFO, SCp,
1948 "New error handler wants BUS reset, cmd %p\n\t", SCp);
1949 scsi_print_command(SCp);
1950
1951 /* In theory, eh_complete should always be null because the
1952 * eh is single threaded, but just in case we're handling a
1953 * reset via sg or something */
1954 spin_lock_irq(SCp->device->host->host_lock);
1955 while (hostdata->eh_complete != NULL) {
1956 spin_unlock_irq(SCp->device->host->host_lock);
1957 msleep_interruptible(100);
1958 spin_lock_irq(SCp->device->host->host_lock);
1959 }
1960
1961 hostdata->eh_complete = &complete;
1962 NCR_700_internal_bus_reset(SCp->device->host);
1963
1964 spin_unlock_irq(SCp->device->host->host_lock);
1965 wait_for_completion(&complete);
1966 spin_lock_irq(SCp->device->host->host_lock);
1967
1968 hostdata->eh_complete = NULL;
1969 /* Revalidate the transport parameters of the failing device */
1970 if(hostdata->fast)
1971 spi_schedule_dv_device(SCp->device);
1972
1973 spin_unlock_irq(SCp->device->host->host_lock);
1974 return SUCCESS;
1975 }
1976
1977 STATIC int
1978 NCR_700_host_reset(struct scsi_cmnd * SCp)
1979 {
1980 scmd_printk(KERN_INFO, SCp, "New error handler wants HOST reset\n\t");
1981 scsi_print_command(SCp);
1982
1983 spin_lock_irq(SCp->device->host->host_lock);
1984
1985 NCR_700_internal_bus_reset(SCp->device->host);
1986 NCR_700_chip_reset(SCp->device->host);
1987
1988 spin_unlock_irq(SCp->device->host->host_lock);
1989
1990 return SUCCESS;
1991 }
1992
1993 STATIC void
1994 NCR_700_set_period(struct scsi_target *STp, int period)
1995 {
1996 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
1997 struct NCR_700_Host_Parameters *hostdata =
1998 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
1999
2000 if(!hostdata->fast)
2001 return;
2002
2003 if(period < hostdata->min_period)
2004 period = hostdata->min_period;
2005
2006 spi_period(STp) = period;
2007 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2008 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2009 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2010 }
2011
2012 STATIC void
2013 NCR_700_set_offset(struct scsi_target *STp, int offset)
2014 {
2015 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
2016 struct NCR_700_Host_Parameters *hostdata =
2017 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
2018 int max_offset = hostdata->chip710
2019 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET;
2020
2021 if(!hostdata->fast)
2022 return;
2023
2024 if(offset > max_offset)
2025 offset = max_offset;
2026
2027 /* if we're currently async, make sure the period is reasonable */
2028 if(spi_offset(STp) == 0 && (spi_period(STp) < hostdata->min_period ||
2029 spi_period(STp) > 0xff))
2030 spi_period(STp) = hostdata->min_period;
2031
2032 spi_offset(STp) = offset;
2033 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2034 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2035 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2036 }
2037
2038
2039
2040 STATIC int
2041 NCR_700_slave_configure(struct scsi_device *SDp)
2042 {
2043 struct NCR_700_Host_Parameters *hostdata =
2044 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2045
2046 SDp->hostdata = kmalloc(GFP_KERNEL, sizeof(struct NCR_700_sense));
2047
2048 if (!SDp->hostdata)
2049 return -ENOMEM;
2050
2051 /* to do here: allocate memory; build a queue_full list */
2052 if(SDp->tagged_supported) {
2053 scsi_set_tag_type(SDp, MSG_ORDERED_TAG);
2054 scsi_activate_tcq(SDp, NCR_700_DEFAULT_TAGS);
2055 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2056 } else {
2057 /* initialise to default depth */
2058 scsi_adjust_queue_depth(SDp, 0, SDp->host->cmd_per_lun);
2059 }
2060 if(hostdata->fast) {
2061 /* Find the correct offset and period via domain validation */
2062 if (!spi_initial_dv(SDp->sdev_target))
2063 spi_dv_device(SDp);
2064 } else {
2065 spi_offset(SDp->sdev_target) = 0;
2066 spi_period(SDp->sdev_target) = 0;
2067 }
2068 return 0;
2069 }
2070
2071 STATIC void
2072 NCR_700_slave_destroy(struct scsi_device *SDp)
2073 {
2074 kfree(SDp->hostdata);
2075 SDp->hostdata = NULL;
2076 }
2077
2078 static int
2079 NCR_700_change_queue_depth(struct scsi_device *SDp, int depth)
2080 {
2081 if (depth > NCR_700_MAX_TAGS)
2082 depth = NCR_700_MAX_TAGS;
2083
2084 scsi_adjust_queue_depth(SDp, scsi_get_tag_type(SDp), depth);
2085 return depth;
2086 }
2087
2088 static int NCR_700_change_queue_type(struct scsi_device *SDp, int tag_type)
2089 {
2090 int change_tag = ((tag_type ==0 && scsi_get_tag_type(SDp) != 0)
2091 || (tag_type != 0 && scsi_get_tag_type(SDp) == 0));
2092 struct NCR_700_Host_Parameters *hostdata =
2093 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2094
2095 scsi_set_tag_type(SDp, tag_type);
2096
2097 /* We have a global (per target) flag to track whether TCQ is
2098 * enabled, so we'll be turning it off for the entire target here.
2099 * our tag algorithm will fail if we mix tagged and untagged commands,
2100 * so quiesce the device before doing this */
2101 if (change_tag)
2102 scsi_target_quiesce(SDp->sdev_target);
2103
2104 if (!tag_type) {
2105 /* shift back to the default unqueued number of commands
2106 * (the user can still raise this) */
2107 scsi_deactivate_tcq(SDp, SDp->host->cmd_per_lun);
2108 hostdata->tag_negotiated &= ~(1 << sdev_id(SDp));
2109 } else {
2110 /* Here, we cleared the negotiation flag above, so this
2111 * will force the driver to renegotiate */
2112 scsi_activate_tcq(SDp, SDp->queue_depth);
2113 if (change_tag)
2114 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2115 }
2116 if (change_tag)
2117 scsi_target_resume(SDp->sdev_target);
2118
2119 return tag_type;
2120 }
2121
2122 static ssize_t
2123 NCR_700_show_active_tags(struct device *dev, struct device_attribute *attr, char *buf)
2124 {
2125 struct scsi_device *SDp = to_scsi_device(dev);
2126
2127 return snprintf(buf, 20, "%d\n", NCR_700_get_depth(SDp));
2128 }
2129
2130 static struct device_attribute NCR_700_active_tags_attr = {
2131 .attr = {
2132 .name = "active_tags",
2133 .mode = S_IRUGO,
2134 },
2135 .show = NCR_700_show_active_tags,
2136 };
2137
2138 STATIC struct device_attribute *NCR_700_dev_attrs[] = {
2139 &NCR_700_active_tags_attr,
2140 NULL,
2141 };
2142
2143 EXPORT_SYMBOL(NCR_700_detect);
2144 EXPORT_SYMBOL(NCR_700_release);
2145 EXPORT_SYMBOL(NCR_700_intr);
2146
2147 static struct spi_function_template NCR_700_transport_functions = {
2148 .set_period = NCR_700_set_period,
2149 .show_period = 1,
2150 .set_offset = NCR_700_set_offset,
2151 .show_offset = 1,
2152 };
2153
2154 static int __init NCR_700_init(void)
2155 {
2156 NCR_700_transport_template = spi_attach_transport(&NCR_700_transport_functions);
2157 if(!NCR_700_transport_template)
2158 return -ENODEV;
2159 return 0;
2160 }
2161
2162 static void __exit NCR_700_exit(void)
2163 {
2164 spi_release_transport(NCR_700_transport_template);
2165 }
2166
2167 module_init(NCR_700_init);
2168 module_exit(NCR_700_exit);
2169