]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - drivers/scsi/am53c974.c
UBUNTU: Ubuntu-4.13.0-45.50
[mirror_ubuntu-artful-kernel.git] / drivers / scsi / am53c974.c
1 /*
2 * AMD am53c974 driver.
3 * Copyright (c) 2014 Hannes Reinecke, SUSE Linux GmbH
4 */
5
6 #include <linux/kernel.h>
7 #include <linux/module.h>
8 #include <linux/init.h>
9 #include <linux/delay.h>
10 #include <linux/pci.h>
11 #include <linux/interrupt.h>
12
13 #include <scsi/scsi_host.h>
14
15 #include "esp_scsi.h"
16
17 #define DRV_MODULE_NAME "am53c974"
18 #define DRV_MODULE_VERSION "1.00"
19
20 static bool am53c974_debug;
21 static bool am53c974_fenab = true;
22
23 #define esp_dma_log(f, a...) \
24 do { \
25 if (am53c974_debug) \
26 shost_printk(KERN_DEBUG, esp->host, f, ##a); \
27 } while (0)
28
29 #define ESP_DMA_CMD 0x10
30 #define ESP_DMA_STC 0x11
31 #define ESP_DMA_SPA 0x12
32 #define ESP_DMA_WBC 0x13
33 #define ESP_DMA_WAC 0x14
34 #define ESP_DMA_STATUS 0x15
35 #define ESP_DMA_SMDLA 0x16
36 #define ESP_DMA_WMAC 0x17
37
38 #define ESP_DMA_CMD_IDLE 0x00
39 #define ESP_DMA_CMD_BLAST 0x01
40 #define ESP_DMA_CMD_ABORT 0x02
41 #define ESP_DMA_CMD_START 0x03
42 #define ESP_DMA_CMD_MASK 0x03
43 #define ESP_DMA_CMD_DIAG 0x04
44 #define ESP_DMA_CMD_MDL 0x10
45 #define ESP_DMA_CMD_INTE_P 0x20
46 #define ESP_DMA_CMD_INTE_D 0x40
47 #define ESP_DMA_CMD_DIR 0x80
48
49 #define ESP_DMA_STAT_PWDN 0x01
50 #define ESP_DMA_STAT_ERROR 0x02
51 #define ESP_DMA_STAT_ABORT 0x04
52 #define ESP_DMA_STAT_DONE 0x08
53 #define ESP_DMA_STAT_SCSIINT 0x10
54 #define ESP_DMA_STAT_BCMPLT 0x20
55
56 /* EEPROM is accessed with 16-bit values */
57 #define DC390_EEPROM_READ 0x80
58 #define DC390_EEPROM_LEN 0x40
59
60 /*
61 * DC390 EEPROM
62 *
63 * 8 * 4 bytes of per-device options
64 * followed by HBA specific options
65 */
66
67 /* Per-device options */
68 #define DC390_EE_MODE1 0x00
69 #define DC390_EE_SPEED 0x01
70
71 /* HBA-specific options */
72 #define DC390_EE_ADAPT_SCSI_ID 0x40
73 #define DC390_EE_MODE2 0x41
74 #define DC390_EE_DELAY 0x42
75 #define DC390_EE_TAG_CMD_NUM 0x43
76
77 #define DC390_EE_MODE1_PARITY_CHK 0x01
78 #define DC390_EE_MODE1_SYNC_NEGO 0x02
79 #define DC390_EE_MODE1_EN_DISC 0x04
80 #define DC390_EE_MODE1_SEND_START 0x08
81 #define DC390_EE_MODE1_TCQ 0x10
82
83 #define DC390_EE_MODE2_MORE_2DRV 0x01
84 #define DC390_EE_MODE2_GREATER_1G 0x02
85 #define DC390_EE_MODE2_RST_SCSI_BUS 0x04
86 #define DC390_EE_MODE2_ACTIVE_NEGATION 0x08
87 #define DC390_EE_MODE2_NO_SEEK 0x10
88 #define DC390_EE_MODE2_LUN_CHECK 0x20
89
90 struct pci_esp_priv {
91 struct esp *esp;
92 u8 dma_status;
93 };
94
95 static void pci_esp_dma_drain(struct esp *esp);
96
97 static inline struct pci_esp_priv *pci_esp_get_priv(struct esp *esp)
98 {
99 struct pci_dev *pdev = esp->dev;
100
101 return pci_get_drvdata(pdev);
102 }
103
104 static void pci_esp_write8(struct esp *esp, u8 val, unsigned long reg)
105 {
106 iowrite8(val, esp->regs + (reg * 4UL));
107 }
108
109 static u8 pci_esp_read8(struct esp *esp, unsigned long reg)
110 {
111 return ioread8(esp->regs + (reg * 4UL));
112 }
113
114 static void pci_esp_write32(struct esp *esp, u32 val, unsigned long reg)
115 {
116 return iowrite32(val, esp->regs + (reg * 4UL));
117 }
118
119 static dma_addr_t pci_esp_map_single(struct esp *esp, void *buf,
120 size_t sz, int dir)
121 {
122 return pci_map_single(esp->dev, buf, sz, dir);
123 }
124
125 static int pci_esp_map_sg(struct esp *esp, struct scatterlist *sg,
126 int num_sg, int dir)
127 {
128 return pci_map_sg(esp->dev, sg, num_sg, dir);
129 }
130
131 static void pci_esp_unmap_single(struct esp *esp, dma_addr_t addr,
132 size_t sz, int dir)
133 {
134 pci_unmap_single(esp->dev, addr, sz, dir);
135 }
136
137 static void pci_esp_unmap_sg(struct esp *esp, struct scatterlist *sg,
138 int num_sg, int dir)
139 {
140 pci_unmap_sg(esp->dev, sg, num_sg, dir);
141 }
142
143 static int pci_esp_irq_pending(struct esp *esp)
144 {
145 struct pci_esp_priv *pep = pci_esp_get_priv(esp);
146
147 pep->dma_status = pci_esp_read8(esp, ESP_DMA_STATUS);
148 esp_dma_log("dma intr dreg[%02x]\n", pep->dma_status);
149
150 if (pep->dma_status & (ESP_DMA_STAT_ERROR |
151 ESP_DMA_STAT_ABORT |
152 ESP_DMA_STAT_DONE |
153 ESP_DMA_STAT_SCSIINT))
154 return 1;
155
156 return 0;
157 }
158
159 static void pci_esp_reset_dma(struct esp *esp)
160 {
161 /* Nothing to do ? */
162 }
163
164 static void pci_esp_dma_drain(struct esp *esp)
165 {
166 u8 resid;
167 int lim = 1000;
168
169
170 if ((esp->sreg & ESP_STAT_PMASK) == ESP_DOP ||
171 (esp->sreg & ESP_STAT_PMASK) == ESP_DIP)
172 /* Data-In or Data-Out, nothing to be done */
173 return;
174
175 while (--lim > 0) {
176 resid = pci_esp_read8(esp, ESP_FFLAGS) & ESP_FF_FBYTES;
177 if (resid <= 1)
178 break;
179 cpu_relax();
180 }
181
182 /*
183 * When there is a residual BCMPLT will never be set
184 * (obviously). But we still have to issue the BLAST
185 * command, otherwise the data will not being transferred.
186 * But we'll never know when the BLAST operation is
187 * finished. So check for some time and give up eventually.
188 */
189 lim = 1000;
190 pci_esp_write8(esp, ESP_DMA_CMD_DIR | ESP_DMA_CMD_BLAST, ESP_DMA_CMD);
191 while (pci_esp_read8(esp, ESP_DMA_STATUS) & ESP_DMA_STAT_BCMPLT) {
192 if (--lim == 0)
193 break;
194 cpu_relax();
195 }
196 pci_esp_write8(esp, ESP_DMA_CMD_DIR | ESP_DMA_CMD_IDLE, ESP_DMA_CMD);
197 esp_dma_log("DMA blast done (%d tries, %d bytes left)\n", lim, resid);
198 /* BLAST residual handling is currently untested */
199 if (WARN_ON_ONCE(resid == 1)) {
200 struct esp_cmd_entry *ent = esp->active_cmd;
201
202 ent->flags |= ESP_CMD_FLAG_RESIDUAL;
203 }
204 }
205
206 static void pci_esp_dma_invalidate(struct esp *esp)
207 {
208 struct pci_esp_priv *pep = pci_esp_get_priv(esp);
209
210 esp_dma_log("invalidate DMA\n");
211
212 pci_esp_write8(esp, ESP_DMA_CMD_IDLE, ESP_DMA_CMD);
213 pep->dma_status = 0;
214 }
215
216 static int pci_esp_dma_error(struct esp *esp)
217 {
218 struct pci_esp_priv *pep = pci_esp_get_priv(esp);
219
220 if (pep->dma_status & ESP_DMA_STAT_ERROR) {
221 u8 dma_cmd = pci_esp_read8(esp, ESP_DMA_CMD);
222
223 if ((dma_cmd & ESP_DMA_CMD_MASK) == ESP_DMA_CMD_START)
224 pci_esp_write8(esp, ESP_DMA_CMD_ABORT, ESP_DMA_CMD);
225
226 return 1;
227 }
228 if (pep->dma_status & ESP_DMA_STAT_ABORT) {
229 pci_esp_write8(esp, ESP_DMA_CMD_IDLE, ESP_DMA_CMD);
230 pep->dma_status = pci_esp_read8(esp, ESP_DMA_CMD);
231 return 1;
232 }
233 return 0;
234 }
235
236 static void pci_esp_send_dma_cmd(struct esp *esp, u32 addr, u32 esp_count,
237 u32 dma_count, int write, u8 cmd)
238 {
239 struct pci_esp_priv *pep = pci_esp_get_priv(esp);
240 u32 val = 0;
241
242 BUG_ON(!(cmd & ESP_CMD_DMA));
243
244 pep->dma_status = 0;
245
246 /* Set DMA engine to IDLE */
247 if (write)
248 /* DMA write direction logic is inverted */
249 val |= ESP_DMA_CMD_DIR;
250 pci_esp_write8(esp, ESP_DMA_CMD_IDLE | val, ESP_DMA_CMD);
251
252 pci_esp_write8(esp, (esp_count >> 0) & 0xff, ESP_TCLOW);
253 pci_esp_write8(esp, (esp_count >> 8) & 0xff, ESP_TCMED);
254 if (esp->config2 & ESP_CONFIG2_FENAB)
255 pci_esp_write8(esp, (esp_count >> 16) & 0xff, ESP_TCHI);
256
257 pci_esp_write32(esp, esp_count, ESP_DMA_STC);
258 pci_esp_write32(esp, addr, ESP_DMA_SPA);
259
260 esp_dma_log("start dma addr[%x] count[%d:%d]\n",
261 addr, esp_count, dma_count);
262
263 scsi_esp_cmd(esp, cmd);
264 /* Send DMA Start command */
265 pci_esp_write8(esp, ESP_DMA_CMD_START | val, ESP_DMA_CMD);
266 }
267
268 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len)
269 {
270 int dma_limit = 16;
271 u32 base, end;
272
273 /*
274 * If CONFIG2_FENAB is set we can
275 * handle up to 24 bit addresses
276 */
277 if (esp->config2 & ESP_CONFIG2_FENAB)
278 dma_limit = 24;
279
280 if (dma_len > (1U << dma_limit))
281 dma_len = (1U << dma_limit);
282
283 /*
284 * Prevent crossing a 24-bit address boundary.
285 */
286 base = dma_addr & ((1U << 24) - 1U);
287 end = base + dma_len;
288 if (end > (1U << 24))
289 end = (1U <<24);
290 dma_len = end - base;
291
292 return dma_len;
293 }
294
295 static const struct esp_driver_ops pci_esp_ops = {
296 .esp_write8 = pci_esp_write8,
297 .esp_read8 = pci_esp_read8,
298 .map_single = pci_esp_map_single,
299 .map_sg = pci_esp_map_sg,
300 .unmap_single = pci_esp_unmap_single,
301 .unmap_sg = pci_esp_unmap_sg,
302 .irq_pending = pci_esp_irq_pending,
303 .reset_dma = pci_esp_reset_dma,
304 .dma_drain = pci_esp_dma_drain,
305 .dma_invalidate = pci_esp_dma_invalidate,
306 .send_dma_cmd = pci_esp_send_dma_cmd,
307 .dma_error = pci_esp_dma_error,
308 .dma_length_limit = pci_esp_dma_length_limit,
309 };
310
311 /*
312 * Read DC-390 eeprom
313 */
314 static void dc390_eeprom_prepare_read(struct pci_dev *pdev, u8 cmd)
315 {
316 u8 carry_flag = 1, j = 0x80, bval;
317 int i;
318
319 for (i = 0; i < 9; i++) {
320 if (carry_flag) {
321 pci_write_config_byte(pdev, 0x80, 0x40);
322 bval = 0xc0;
323 } else
324 bval = 0x80;
325
326 udelay(160);
327 pci_write_config_byte(pdev, 0x80, bval);
328 udelay(160);
329 pci_write_config_byte(pdev, 0x80, 0);
330 udelay(160);
331
332 carry_flag = (cmd & j) ? 1 : 0;
333 j >>= 1;
334 }
335 }
336
337 static u16 dc390_eeprom_get_data(struct pci_dev *pdev)
338 {
339 int i;
340 u16 wval = 0;
341 u8 bval;
342
343 for (i = 0; i < 16; i++) {
344 wval <<= 1;
345
346 pci_write_config_byte(pdev, 0x80, 0x80);
347 udelay(160);
348 pci_write_config_byte(pdev, 0x80, 0x40);
349 udelay(160);
350 pci_read_config_byte(pdev, 0x00, &bval);
351
352 if (bval == 0x22)
353 wval |= 1;
354 }
355
356 return wval;
357 }
358
359 static void dc390_read_eeprom(struct pci_dev *pdev, u16 *ptr)
360 {
361 u8 cmd = DC390_EEPROM_READ, i;
362
363 for (i = 0; i < DC390_EEPROM_LEN; i++) {
364 pci_write_config_byte(pdev, 0xc0, 0);
365 udelay(160);
366
367 dc390_eeprom_prepare_read(pdev, cmd++);
368 *ptr++ = dc390_eeprom_get_data(pdev);
369
370 pci_write_config_byte(pdev, 0x80, 0);
371 pci_write_config_byte(pdev, 0x80, 0);
372 udelay(160);
373 }
374 }
375
376 static void dc390_check_eeprom(struct esp *esp)
377 {
378 u8 EEbuf[128];
379 u16 *ptr = (u16 *)EEbuf, wval = 0;
380 int i;
381
382 dc390_read_eeprom((struct pci_dev *)esp->dev, ptr);
383
384 for (i = 0; i < DC390_EEPROM_LEN; i++, ptr++)
385 wval += *ptr;
386
387 /* no Tekram EEprom found */
388 if (wval != 0x1234) {
389 struct pci_dev *pdev = esp->dev;
390 dev_printk(KERN_INFO, &pdev->dev,
391 "No valid Tekram EEprom found\n");
392 return;
393 }
394 esp->scsi_id = EEbuf[DC390_EE_ADAPT_SCSI_ID];
395 esp->num_tags = 2 << EEbuf[DC390_EE_TAG_CMD_NUM];
396 if (EEbuf[DC390_EE_MODE2] & DC390_EE_MODE2_ACTIVE_NEGATION)
397 esp->config4 |= ESP_CONFIG4_RADE | ESP_CONFIG4_RAE;
398 }
399
400 static int pci_esp_probe_one(struct pci_dev *pdev,
401 const struct pci_device_id *id)
402 {
403 struct scsi_host_template *hostt = &scsi_esp_template;
404 int err = -ENODEV;
405 struct Scsi_Host *shost;
406 struct esp *esp;
407 struct pci_esp_priv *pep;
408
409 if (pci_enable_device(pdev)) {
410 dev_printk(KERN_INFO, &pdev->dev, "cannot enable device\n");
411 return -ENODEV;
412 }
413
414 if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
415 dev_printk(KERN_INFO, &pdev->dev,
416 "failed to set 32bit DMA mask\n");
417 goto fail_disable_device;
418 }
419
420 shost = scsi_host_alloc(hostt, sizeof(struct esp));
421 if (!shost) {
422 dev_printk(KERN_INFO, &pdev->dev,
423 "failed to allocate scsi host\n");
424 err = -ENOMEM;
425 goto fail_disable_device;
426 }
427
428 pep = kzalloc(sizeof(struct pci_esp_priv), GFP_KERNEL);
429 if (!pep) {
430 dev_printk(KERN_INFO, &pdev->dev,
431 "failed to allocate esp_priv\n");
432 err = -ENOMEM;
433 goto fail_host_alloc;
434 }
435
436 esp = shost_priv(shost);
437 esp->host = shost;
438 esp->dev = pdev;
439 esp->ops = &pci_esp_ops;
440 /*
441 * The am53c974 HBA has a design flaw of generating
442 * spurious DMA completion interrupts when using
443 * DMA for command submission.
444 */
445 esp->flags |= ESP_FLAG_USE_FIFO;
446 /*
447 * Enable CONFIG2_FENAB to allow for large DMA transfers
448 */
449 if (am53c974_fenab)
450 esp->config2 |= ESP_CONFIG2_FENAB;
451
452 pep->esp = esp;
453
454 if (pci_request_regions(pdev, DRV_MODULE_NAME)) {
455 dev_printk(KERN_ERR, &pdev->dev,
456 "pci memory selection failed\n");
457 goto fail_priv_alloc;
458 }
459
460 esp->regs = pci_iomap(pdev, 0, pci_resource_len(pdev, 0));
461 if (!esp->regs) {
462 dev_printk(KERN_ERR, &pdev->dev, "pci I/O map failed\n");
463 err = -EINVAL;
464 goto fail_release_regions;
465 }
466 esp->dma_regs = esp->regs;
467
468 pci_set_master(pdev);
469
470 esp->command_block = pci_alloc_consistent(pdev, 16,
471 &esp->command_block_dma);
472 if (!esp->command_block) {
473 dev_printk(KERN_ERR, &pdev->dev,
474 "failed to allocate command block\n");
475 err = -ENOMEM;
476 goto fail_unmap_regs;
477 }
478
479 pci_set_drvdata(pdev, pep);
480
481 err = request_irq(pdev->irq, scsi_esp_intr, IRQF_SHARED,
482 DRV_MODULE_NAME, esp);
483 if (err < 0) {
484 dev_printk(KERN_ERR, &pdev->dev, "failed to register IRQ\n");
485 goto fail_unmap_command_block;
486 }
487
488 esp->scsi_id = 7;
489 dc390_check_eeprom(esp);
490
491 shost->this_id = esp->scsi_id;
492 shost->max_id = 8;
493 shost->irq = pdev->irq;
494 shost->io_port = pci_resource_start(pdev, 0);
495 shost->n_io_port = pci_resource_len(pdev, 0);
496 shost->unique_id = shost->io_port;
497 esp->scsi_id_mask = (1 << esp->scsi_id);
498 /* Assume 40MHz clock */
499 esp->cfreq = 40000000;
500
501 err = scsi_esp_register(esp, &pdev->dev);
502 if (err)
503 goto fail_free_irq;
504
505 return 0;
506
507 fail_free_irq:
508 free_irq(pdev->irq, esp);
509 fail_unmap_command_block:
510 pci_set_drvdata(pdev, NULL);
511 pci_free_consistent(pdev, 16, esp->command_block,
512 esp->command_block_dma);
513 fail_unmap_regs:
514 pci_iounmap(pdev, esp->regs);
515 fail_release_regions:
516 pci_release_regions(pdev);
517 fail_priv_alloc:
518 kfree(pep);
519 fail_host_alloc:
520 scsi_host_put(shost);
521 fail_disable_device:
522 pci_disable_device(pdev);
523
524 return err;
525 }
526
527 static void pci_esp_remove_one(struct pci_dev *pdev)
528 {
529 struct pci_esp_priv *pep = pci_get_drvdata(pdev);
530 struct esp *esp = pep->esp;
531
532 scsi_esp_unregister(esp);
533 free_irq(pdev->irq, esp);
534 pci_set_drvdata(pdev, NULL);
535 pci_free_consistent(pdev, 16, esp->command_block,
536 esp->command_block_dma);
537 pci_iounmap(pdev, esp->regs);
538 pci_release_regions(pdev);
539 pci_disable_device(pdev);
540 kfree(pep);
541
542 scsi_host_put(esp->host);
543 }
544
545 static struct pci_device_id am53c974_pci_tbl[] = {
546 { PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_SCSI,
547 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 },
548 { }
549 };
550 MODULE_DEVICE_TABLE(pci, am53c974_pci_tbl);
551
552 static struct pci_driver am53c974_driver = {
553 .name = DRV_MODULE_NAME,
554 .id_table = am53c974_pci_tbl,
555 .probe = pci_esp_probe_one,
556 .remove = pci_esp_remove_one,
557 };
558
559 static int __init am53c974_module_init(void)
560 {
561 return pci_register_driver(&am53c974_driver);
562 }
563
564 static void __exit am53c974_module_exit(void)
565 {
566 pci_unregister_driver(&am53c974_driver);
567 }
568
569 MODULE_DESCRIPTION("AM53C974 SCSI driver");
570 MODULE_AUTHOR("Hannes Reinecke <hare@suse.de>");
571 MODULE_LICENSE("GPL");
572 MODULE_VERSION(DRV_MODULE_VERSION);
573 MODULE_ALIAS("tmscsim");
574
575 module_param(am53c974_debug, bool, 0644);
576 MODULE_PARM_DESC(am53c974_debug, "Enable debugging");
577
578 module_param(am53c974_fenab, bool, 0444);
579 MODULE_PARM_DESC(am53c974_fenab, "Enable 24-bit DMA transfer sizes");
580
581 module_init(am53c974_module_init);
582 module_exit(am53c974_module_exit);