]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * sata_via.c - VIA Serial ATA controllers | |
3 | * | |
4 | * Maintained by: Tejun Heo <tj@kernel.org> | |
5 | * Please ALWAYS copy linux-ide@vger.kernel.org | |
6 | * on emails. | |
7 | * | |
8 | * Copyright 2003-2004 Red Hat, Inc. All rights reserved. | |
9 | * Copyright 2003-2004 Jeff Garzik | |
10 | * | |
11 | * | |
12 | * This program is free software; you can redistribute it and/or modify | |
13 | * it under the terms of the GNU General Public License as published by | |
14 | * the Free Software Foundation; either version 2, or (at your option) | |
15 | * any later version. | |
16 | * | |
17 | * This program is distributed in the hope that it will be useful, | |
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
20 | * GNU General Public License for more details. | |
21 | * | |
22 | * You should have received a copy of the GNU General Public License | |
23 | * along with this program; see the file COPYING. If not, write to | |
24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | |
25 | * | |
26 | * | |
27 | * libata documentation is available via 'make {ps|pdf}docs', | |
28 | * as Documentation/DocBook/libata.* | |
29 | * | |
30 | * Hardware documentation available under NDA. | |
31 | * | |
32 | * | |
33 | * | |
34 | */ | |
35 | ||
36 | #include <linux/kernel.h> | |
37 | #include <linux/module.h> | |
38 | #include <linux/pci.h> | |
39 | #include <linux/blkdev.h> | |
40 | #include <linux/delay.h> | |
41 | #include <linux/device.h> | |
42 | #include <scsi/scsi.h> | |
43 | #include <scsi/scsi_cmnd.h> | |
44 | #include <scsi/scsi_host.h> | |
45 | #include <linux/libata.h> | |
46 | ||
47 | #define DRV_NAME "sata_via" | |
48 | #define DRV_VERSION "2.6" | |
49 | ||
50 | /* | |
51 | * vt8251 is different from other sata controllers of VIA. It has two | |
52 | * channels, each channel has both Master and Slave slot. | |
53 | */ | |
54 | enum board_ids_enum { | |
55 | vt6420, | |
56 | vt6421, | |
57 | vt8251, | |
58 | }; | |
59 | ||
60 | enum { | |
61 | SATA_CHAN_ENAB = 0x40, /* SATA channel enable */ | |
62 | SATA_INT_GATE = 0x41, /* SATA interrupt gating */ | |
63 | SATA_NATIVE_MODE = 0x42, /* Native mode enable */ | |
64 | SVIA_MISC_3 = 0x46, /* Miscellaneous Control III */ | |
65 | PATA_UDMA_TIMING = 0xB3, /* PATA timing for DMA/ cable detect */ | |
66 | PATA_PIO_TIMING = 0xAB, /* PATA timing register */ | |
67 | ||
68 | PORT0 = (1 << 1), | |
69 | PORT1 = (1 << 0), | |
70 | ALL_PORTS = PORT0 | PORT1, | |
71 | ||
72 | NATIVE_MODE_ALL = (1 << 7) | (1 << 6) | (1 << 5) | (1 << 4), | |
73 | ||
74 | SATA_EXT_PHY = (1 << 6), /* 0==use PATA, 1==ext phy */ | |
75 | ||
76 | SATA_HOTPLUG = (1 << 5), /* enable IRQ on hotplug */ | |
77 | }; | |
78 | ||
79 | struct svia_priv { | |
80 | bool wd_workaround; | |
81 | }; | |
82 | ||
83 | static int svia_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); | |
84 | #ifdef CONFIG_PM_SLEEP | |
85 | static int svia_pci_device_resume(struct pci_dev *pdev); | |
86 | #endif | |
87 | static int svia_scr_read(struct ata_link *link, unsigned int sc_reg, u32 *val); | |
88 | static int svia_scr_write(struct ata_link *link, unsigned int sc_reg, u32 val); | |
89 | static int vt8251_scr_read(struct ata_link *link, unsigned int scr, u32 *val); | |
90 | static int vt8251_scr_write(struct ata_link *link, unsigned int scr, u32 val); | |
91 | static void svia_tf_load(struct ata_port *ap, const struct ata_taskfile *tf); | |
92 | static void svia_noop_freeze(struct ata_port *ap); | |
93 | static int vt6420_prereset(struct ata_link *link, unsigned long deadline); | |
94 | static void vt6420_bmdma_start(struct ata_queued_cmd *qc); | |
95 | static int vt6421_pata_cable_detect(struct ata_port *ap); | |
96 | static void vt6421_set_pio_mode(struct ata_port *ap, struct ata_device *adev); | |
97 | static void vt6421_set_dma_mode(struct ata_port *ap, struct ata_device *adev); | |
98 | static void vt6421_error_handler(struct ata_port *ap); | |
99 | ||
100 | static const struct pci_device_id svia_pci_tbl[] = { | |
101 | { PCI_VDEVICE(VIA, 0x5337), vt6420 }, | |
102 | { PCI_VDEVICE(VIA, 0x0591), vt6420 }, /* 2 sata chnls (Master) */ | |
103 | { PCI_VDEVICE(VIA, 0x3149), vt6420 }, /* 2 sata chnls (Master) */ | |
104 | { PCI_VDEVICE(VIA, 0x3249), vt6421 }, /* 2 sata chnls, 1 pata chnl */ | |
105 | { PCI_VDEVICE(VIA, 0x5372), vt6420 }, | |
106 | { PCI_VDEVICE(VIA, 0x7372), vt6420 }, | |
107 | { PCI_VDEVICE(VIA, 0x5287), vt8251 }, /* 2 sata chnls (Master/Slave) */ | |
108 | { PCI_VDEVICE(VIA, 0x9000), vt8251 }, | |
109 | ||
110 | { } /* terminate list */ | |
111 | }; | |
112 | ||
113 | static struct pci_driver svia_pci_driver = { | |
114 | .name = DRV_NAME, | |
115 | .id_table = svia_pci_tbl, | |
116 | .probe = svia_init_one, | |
117 | #ifdef CONFIG_PM_SLEEP | |
118 | .suspend = ata_pci_device_suspend, | |
119 | .resume = svia_pci_device_resume, | |
120 | #endif | |
121 | .remove = ata_pci_remove_one, | |
122 | }; | |
123 | ||
124 | static struct scsi_host_template svia_sht = { | |
125 | ATA_BMDMA_SHT(DRV_NAME), | |
126 | }; | |
127 | ||
128 | static struct ata_port_operations svia_base_ops = { | |
129 | .inherits = &ata_bmdma_port_ops, | |
130 | .sff_tf_load = svia_tf_load, | |
131 | }; | |
132 | ||
133 | static struct ata_port_operations vt6420_sata_ops = { | |
134 | .inherits = &svia_base_ops, | |
135 | .freeze = svia_noop_freeze, | |
136 | .prereset = vt6420_prereset, | |
137 | .bmdma_start = vt6420_bmdma_start, | |
138 | }; | |
139 | ||
140 | static struct ata_port_operations vt6421_pata_ops = { | |
141 | .inherits = &svia_base_ops, | |
142 | .cable_detect = vt6421_pata_cable_detect, | |
143 | .set_piomode = vt6421_set_pio_mode, | |
144 | .set_dmamode = vt6421_set_dma_mode, | |
145 | }; | |
146 | ||
147 | static struct ata_port_operations vt6421_sata_ops = { | |
148 | .inherits = &svia_base_ops, | |
149 | .scr_read = svia_scr_read, | |
150 | .scr_write = svia_scr_write, | |
151 | .error_handler = vt6421_error_handler, | |
152 | }; | |
153 | ||
154 | static struct ata_port_operations vt8251_ops = { | |
155 | .inherits = &svia_base_ops, | |
156 | .hardreset = sata_std_hardreset, | |
157 | .scr_read = vt8251_scr_read, | |
158 | .scr_write = vt8251_scr_write, | |
159 | }; | |
160 | ||
161 | static const struct ata_port_info vt6420_port_info = { | |
162 | .flags = ATA_FLAG_SATA, | |
163 | .pio_mask = ATA_PIO4, | |
164 | .mwdma_mask = ATA_MWDMA2, | |
165 | .udma_mask = ATA_UDMA6, | |
166 | .port_ops = &vt6420_sata_ops, | |
167 | }; | |
168 | ||
169 | static struct ata_port_info vt6421_sport_info = { | |
170 | .flags = ATA_FLAG_SATA, | |
171 | .pio_mask = ATA_PIO4, | |
172 | .mwdma_mask = ATA_MWDMA2, | |
173 | .udma_mask = ATA_UDMA6, | |
174 | .port_ops = &vt6421_sata_ops, | |
175 | }; | |
176 | ||
177 | static struct ata_port_info vt6421_pport_info = { | |
178 | .flags = ATA_FLAG_SLAVE_POSS, | |
179 | .pio_mask = ATA_PIO4, | |
180 | /* No MWDMA */ | |
181 | .udma_mask = ATA_UDMA6, | |
182 | .port_ops = &vt6421_pata_ops, | |
183 | }; | |
184 | ||
185 | static struct ata_port_info vt8251_port_info = { | |
186 | .flags = ATA_FLAG_SATA | ATA_FLAG_SLAVE_POSS, | |
187 | .pio_mask = ATA_PIO4, | |
188 | .mwdma_mask = ATA_MWDMA2, | |
189 | .udma_mask = ATA_UDMA6, | |
190 | .port_ops = &vt8251_ops, | |
191 | }; | |
192 | ||
193 | MODULE_AUTHOR("Jeff Garzik"); | |
194 | MODULE_DESCRIPTION("SCSI low-level driver for VIA SATA controllers"); | |
195 | MODULE_LICENSE("GPL"); | |
196 | MODULE_DEVICE_TABLE(pci, svia_pci_tbl); | |
197 | MODULE_VERSION(DRV_VERSION); | |
198 | ||
199 | static int svia_scr_read(struct ata_link *link, unsigned int sc_reg, u32 *val) | |
200 | { | |
201 | if (sc_reg > SCR_CONTROL) | |
202 | return -EINVAL; | |
203 | *val = ioread32(link->ap->ioaddr.scr_addr + (4 * sc_reg)); | |
204 | return 0; | |
205 | } | |
206 | ||
207 | static int svia_scr_write(struct ata_link *link, unsigned int sc_reg, u32 val) | |
208 | { | |
209 | if (sc_reg > SCR_CONTROL) | |
210 | return -EINVAL; | |
211 | iowrite32(val, link->ap->ioaddr.scr_addr + (4 * sc_reg)); | |
212 | return 0; | |
213 | } | |
214 | ||
215 | static int vt8251_scr_read(struct ata_link *link, unsigned int scr, u32 *val) | |
216 | { | |
217 | static const u8 ipm_tbl[] = { 1, 2, 6, 0 }; | |
218 | struct pci_dev *pdev = to_pci_dev(link->ap->host->dev); | |
219 | int slot = 2 * link->ap->port_no + link->pmp; | |
220 | u32 v = 0; | |
221 | u8 raw; | |
222 | ||
223 | switch (scr) { | |
224 | case SCR_STATUS: | |
225 | pci_read_config_byte(pdev, 0xA0 + slot, &raw); | |
226 | ||
227 | /* read the DET field, bit0 and 1 of the config byte */ | |
228 | v |= raw & 0x03; | |
229 | ||
230 | /* read the SPD field, bit4 of the configure byte */ | |
231 | if (raw & (1 << 4)) | |
232 | v |= 0x02 << 4; | |
233 | else | |
234 | v |= 0x01 << 4; | |
235 | ||
236 | /* read the IPM field, bit2 and 3 of the config byte */ | |
237 | v |= ipm_tbl[(raw >> 2) & 0x3]; | |
238 | break; | |
239 | ||
240 | case SCR_ERROR: | |
241 | /* devices other than 5287 uses 0xA8 as base */ | |
242 | WARN_ON(pdev->device != 0x5287); | |
243 | pci_read_config_dword(pdev, 0xB0 + slot * 4, &v); | |
244 | break; | |
245 | ||
246 | case SCR_CONTROL: | |
247 | pci_read_config_byte(pdev, 0xA4 + slot, &raw); | |
248 | ||
249 | /* read the DET field, bit0 and bit1 */ | |
250 | v |= ((raw & 0x02) << 1) | (raw & 0x01); | |
251 | ||
252 | /* read the IPM field, bit2 and bit3 */ | |
253 | v |= ((raw >> 2) & 0x03) << 8; | |
254 | break; | |
255 | ||
256 | default: | |
257 | return -EINVAL; | |
258 | } | |
259 | ||
260 | *val = v; | |
261 | return 0; | |
262 | } | |
263 | ||
264 | static int vt8251_scr_write(struct ata_link *link, unsigned int scr, u32 val) | |
265 | { | |
266 | struct pci_dev *pdev = to_pci_dev(link->ap->host->dev); | |
267 | int slot = 2 * link->ap->port_no + link->pmp; | |
268 | u32 v = 0; | |
269 | ||
270 | switch (scr) { | |
271 | case SCR_ERROR: | |
272 | /* devices other than 5287 uses 0xA8 as base */ | |
273 | WARN_ON(pdev->device != 0x5287); | |
274 | pci_write_config_dword(pdev, 0xB0 + slot * 4, val); | |
275 | return 0; | |
276 | ||
277 | case SCR_CONTROL: | |
278 | /* set the DET field */ | |
279 | v |= ((val & 0x4) >> 1) | (val & 0x1); | |
280 | ||
281 | /* set the IPM field */ | |
282 | v |= ((val >> 8) & 0x3) << 2; | |
283 | ||
284 | pci_write_config_byte(pdev, 0xA4 + slot, v); | |
285 | return 0; | |
286 | ||
287 | default: | |
288 | return -EINVAL; | |
289 | } | |
290 | } | |
291 | ||
292 | /** | |
293 | * svia_tf_load - send taskfile registers to host controller | |
294 | * @ap: Port to which output is sent | |
295 | * @tf: ATA taskfile register set | |
296 | * | |
297 | * Outputs ATA taskfile to standard ATA host controller. | |
298 | * | |
299 | * This is to fix the internal bug of via chipsets, which will | |
300 | * reset the device register after changing the IEN bit on ctl | |
301 | * register. | |
302 | */ | |
303 | static void svia_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) | |
304 | { | |
305 | struct ata_taskfile ttf; | |
306 | ||
307 | if (tf->ctl != ap->last_ctl) { | |
308 | ttf = *tf; | |
309 | ttf.flags |= ATA_TFLAG_DEVICE; | |
310 | tf = &ttf; | |
311 | } | |
312 | ata_sff_tf_load(ap, tf); | |
313 | } | |
314 | ||
315 | static void svia_noop_freeze(struct ata_port *ap) | |
316 | { | |
317 | /* Some VIA controllers choke if ATA_NIEN is manipulated in | |
318 | * certain way. Leave it alone and just clear pending IRQ. | |
319 | */ | |
320 | ap->ops->sff_check_status(ap); | |
321 | ata_bmdma_irq_clear(ap); | |
322 | } | |
323 | ||
324 | /** | |
325 | * vt6420_prereset - prereset for vt6420 | |
326 | * @link: target ATA link | |
327 | * @deadline: deadline jiffies for the operation | |
328 | * | |
329 | * SCR registers on vt6420 are pieces of shit and may hang the | |
330 | * whole machine completely if accessed with the wrong timing. | |
331 | * To avoid such catastrophe, vt6420 doesn't provide generic SCR | |
332 | * access operations, but uses SStatus and SControl only during | |
333 | * boot probing in controlled way. | |
334 | * | |
335 | * As the old (pre EH update) probing code is proven to work, we | |
336 | * strictly follow the access pattern. | |
337 | * | |
338 | * LOCKING: | |
339 | * Kernel thread context (may sleep) | |
340 | * | |
341 | * RETURNS: | |
342 | * 0 on success, -errno otherwise. | |
343 | */ | |
344 | static int vt6420_prereset(struct ata_link *link, unsigned long deadline) | |
345 | { | |
346 | struct ata_port *ap = link->ap; | |
347 | struct ata_eh_context *ehc = &ap->link.eh_context; | |
348 | unsigned long timeout = jiffies + (HZ * 5); | |
349 | u32 sstatus, scontrol; | |
350 | int online; | |
351 | ||
352 | /* don't do any SCR stuff if we're not loading */ | |
353 | if (!(ap->pflags & ATA_PFLAG_LOADING)) | |
354 | goto skip_scr; | |
355 | ||
356 | /* Resume phy. This is the old SATA resume sequence */ | |
357 | svia_scr_write(link, SCR_CONTROL, 0x300); | |
358 | svia_scr_read(link, SCR_CONTROL, &scontrol); /* flush */ | |
359 | ||
360 | /* wait for phy to become ready, if necessary */ | |
361 | do { | |
362 | ata_msleep(link->ap, 200); | |
363 | svia_scr_read(link, SCR_STATUS, &sstatus); | |
364 | if ((sstatus & 0xf) != 1) | |
365 | break; | |
366 | } while (time_before(jiffies, timeout)); | |
367 | ||
368 | /* open code sata_print_link_status() */ | |
369 | svia_scr_read(link, SCR_STATUS, &sstatus); | |
370 | svia_scr_read(link, SCR_CONTROL, &scontrol); | |
371 | ||
372 | online = (sstatus & 0xf) == 0x3; | |
373 | ||
374 | ata_port_info(ap, | |
375 | "SATA link %s 1.5 Gbps (SStatus %X SControl %X)\n", | |
376 | online ? "up" : "down", sstatus, scontrol); | |
377 | ||
378 | /* SStatus is read one more time */ | |
379 | svia_scr_read(link, SCR_STATUS, &sstatus); | |
380 | ||
381 | if (!online) { | |
382 | /* tell EH to bail */ | |
383 | ehc->i.action &= ~ATA_EH_RESET; | |
384 | return 0; | |
385 | } | |
386 | ||
387 | skip_scr: | |
388 | /* wait for !BSY */ | |
389 | ata_sff_wait_ready(link, deadline); | |
390 | ||
391 | return 0; | |
392 | } | |
393 | ||
394 | static void vt6420_bmdma_start(struct ata_queued_cmd *qc) | |
395 | { | |
396 | struct ata_port *ap = qc->ap; | |
397 | if ((qc->tf.command == ATA_CMD_PACKET) && | |
398 | (qc->scsicmd->sc_data_direction == DMA_TO_DEVICE)) { | |
399 | /* Prevents corruption on some ATAPI burners */ | |
400 | ata_sff_pause(ap); | |
401 | } | |
402 | ata_bmdma_start(qc); | |
403 | } | |
404 | ||
405 | static int vt6421_pata_cable_detect(struct ata_port *ap) | |
406 | { | |
407 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | |
408 | u8 tmp; | |
409 | ||
410 | pci_read_config_byte(pdev, PATA_UDMA_TIMING, &tmp); | |
411 | if (tmp & 0x10) | |
412 | return ATA_CBL_PATA40; | |
413 | return ATA_CBL_PATA80; | |
414 | } | |
415 | ||
416 | static void vt6421_set_pio_mode(struct ata_port *ap, struct ata_device *adev) | |
417 | { | |
418 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | |
419 | static const u8 pio_bits[] = { 0xA8, 0x65, 0x65, 0x31, 0x20 }; | |
420 | pci_write_config_byte(pdev, PATA_PIO_TIMING - adev->devno, | |
421 | pio_bits[adev->pio_mode - XFER_PIO_0]); | |
422 | } | |
423 | ||
424 | static void vt6421_set_dma_mode(struct ata_port *ap, struct ata_device *adev) | |
425 | { | |
426 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | |
427 | static const u8 udma_bits[] = { 0xEE, 0xE8, 0xE6, 0xE4, 0xE2, 0xE1, 0xE0, 0xE0 }; | |
428 | pci_write_config_byte(pdev, PATA_UDMA_TIMING - adev->devno, | |
429 | udma_bits[adev->dma_mode - XFER_UDMA_0]); | |
430 | } | |
431 | ||
432 | static const unsigned int svia_bar_sizes[] = { | |
433 | 8, 4, 8, 4, 16, 256 | |
434 | }; | |
435 | ||
436 | static const unsigned int vt6421_bar_sizes[] = { | |
437 | 16, 16, 16, 16, 32, 128 | |
438 | }; | |
439 | ||
440 | static void __iomem *svia_scr_addr(void __iomem *addr, unsigned int port) | |
441 | { | |
442 | return addr + (port * 128); | |
443 | } | |
444 | ||
445 | static void __iomem *vt6421_scr_addr(void __iomem *addr, unsigned int port) | |
446 | { | |
447 | return addr + (port * 64); | |
448 | } | |
449 | ||
450 | static void vt6421_init_addrs(struct ata_port *ap) | |
451 | { | |
452 | void __iomem * const * iomap = ap->host->iomap; | |
453 | void __iomem *reg_addr = iomap[ap->port_no]; | |
454 | void __iomem *bmdma_addr = iomap[4] + (ap->port_no * 8); | |
455 | struct ata_ioports *ioaddr = &ap->ioaddr; | |
456 | ||
457 | ioaddr->cmd_addr = reg_addr; | |
458 | ioaddr->altstatus_addr = | |
459 | ioaddr->ctl_addr = (void __iomem *) | |
460 | ((unsigned long)(reg_addr + 8) | ATA_PCI_CTL_OFS); | |
461 | ioaddr->bmdma_addr = bmdma_addr; | |
462 | ioaddr->scr_addr = vt6421_scr_addr(iomap[5], ap->port_no); | |
463 | ||
464 | ata_sff_std_ports(ioaddr); | |
465 | ||
466 | ata_port_pbar_desc(ap, ap->port_no, -1, "port"); | |
467 | ata_port_pbar_desc(ap, 4, ap->port_no * 8, "bmdma"); | |
468 | } | |
469 | ||
470 | static int vt6420_prepare_host(struct pci_dev *pdev, struct ata_host **r_host) | |
471 | { | |
472 | const struct ata_port_info *ppi[] = { &vt6420_port_info, NULL }; | |
473 | struct ata_host *host; | |
474 | int rc; | |
475 | ||
476 | rc = ata_pci_bmdma_prepare_host(pdev, ppi, &host); | |
477 | if (rc) | |
478 | return rc; | |
479 | *r_host = host; | |
480 | ||
481 | rc = pcim_iomap_regions(pdev, 1 << 5, DRV_NAME); | |
482 | if (rc) { | |
483 | dev_err(&pdev->dev, "failed to iomap PCI BAR 5\n"); | |
484 | return rc; | |
485 | } | |
486 | ||
487 | host->ports[0]->ioaddr.scr_addr = svia_scr_addr(host->iomap[5], 0); | |
488 | host->ports[1]->ioaddr.scr_addr = svia_scr_addr(host->iomap[5], 1); | |
489 | ||
490 | return 0; | |
491 | } | |
492 | ||
493 | static int vt6421_prepare_host(struct pci_dev *pdev, struct ata_host **r_host) | |
494 | { | |
495 | const struct ata_port_info *ppi[] = | |
496 | { &vt6421_sport_info, &vt6421_sport_info, &vt6421_pport_info }; | |
497 | struct ata_host *host; | |
498 | int i, rc; | |
499 | ||
500 | *r_host = host = ata_host_alloc_pinfo(&pdev->dev, ppi, ARRAY_SIZE(ppi)); | |
501 | if (!host) { | |
502 | dev_err(&pdev->dev, "failed to allocate host\n"); | |
503 | return -ENOMEM; | |
504 | } | |
505 | ||
506 | rc = pcim_iomap_regions(pdev, 0x3f, DRV_NAME); | |
507 | if (rc) { | |
508 | dev_err(&pdev->dev, "failed to request/iomap PCI BARs (errno=%d)\n", | |
509 | rc); | |
510 | return rc; | |
511 | } | |
512 | host->iomap = pcim_iomap_table(pdev); | |
513 | ||
514 | for (i = 0; i < host->n_ports; i++) | |
515 | vt6421_init_addrs(host->ports[i]); | |
516 | ||
517 | rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK); | |
518 | if (rc) | |
519 | return rc; | |
520 | rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK); | |
521 | if (rc) | |
522 | return rc; | |
523 | ||
524 | return 0; | |
525 | } | |
526 | ||
527 | static int vt8251_prepare_host(struct pci_dev *pdev, struct ata_host **r_host) | |
528 | { | |
529 | const struct ata_port_info *ppi[] = { &vt8251_port_info, NULL }; | |
530 | struct ata_host *host; | |
531 | int i, rc; | |
532 | ||
533 | rc = ata_pci_bmdma_prepare_host(pdev, ppi, &host); | |
534 | if (rc) | |
535 | return rc; | |
536 | *r_host = host; | |
537 | ||
538 | rc = pcim_iomap_regions(pdev, 1 << 5, DRV_NAME); | |
539 | if (rc) { | |
540 | dev_err(&pdev->dev, "failed to iomap PCI BAR 5\n"); | |
541 | return rc; | |
542 | } | |
543 | ||
544 | /* 8251 hosts four sata ports as M/S of the two channels */ | |
545 | for (i = 0; i < host->n_ports; i++) | |
546 | ata_slave_link_init(host->ports[i]); | |
547 | ||
548 | return 0; | |
549 | } | |
550 | ||
551 | static void svia_wd_fix(struct pci_dev *pdev) | |
552 | { | |
553 | u8 tmp8; | |
554 | ||
555 | pci_read_config_byte(pdev, 0x52, &tmp8); | |
556 | pci_write_config_byte(pdev, 0x52, tmp8 | BIT(2)); | |
557 | } | |
558 | ||
559 | static irqreturn_t vt6421_interrupt(int irq, void *dev_instance) | |
560 | { | |
561 | struct ata_host *host = dev_instance; | |
562 | irqreturn_t rc = ata_bmdma_interrupt(irq, dev_instance); | |
563 | ||
564 | /* if the IRQ was not handled, it might be a hotplug IRQ */ | |
565 | if (rc != IRQ_HANDLED) { | |
566 | u32 serror; | |
567 | unsigned long flags; | |
568 | ||
569 | spin_lock_irqsave(&host->lock, flags); | |
570 | /* check for hotplug on port 0 */ | |
571 | svia_scr_read(&host->ports[0]->link, SCR_ERROR, &serror); | |
572 | if (serror & SERR_PHYRDY_CHG) { | |
573 | ata_ehi_hotplugged(&host->ports[0]->link.eh_info); | |
574 | ata_port_freeze(host->ports[0]); | |
575 | rc = IRQ_HANDLED; | |
576 | } | |
577 | /* check for hotplug on port 1 */ | |
578 | svia_scr_read(&host->ports[1]->link, SCR_ERROR, &serror); | |
579 | if (serror & SERR_PHYRDY_CHG) { | |
580 | ata_ehi_hotplugged(&host->ports[1]->link.eh_info); | |
581 | ata_port_freeze(host->ports[1]); | |
582 | rc = IRQ_HANDLED; | |
583 | } | |
584 | spin_unlock_irqrestore(&host->lock, flags); | |
585 | } | |
586 | ||
587 | return rc; | |
588 | } | |
589 | ||
590 | static void vt6421_error_handler(struct ata_port *ap) | |
591 | { | |
592 | struct svia_priv *hpriv = ap->host->private_data; | |
593 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | |
594 | u32 serror; | |
595 | ||
596 | /* see svia_configure() for description */ | |
597 | if (!hpriv->wd_workaround) { | |
598 | svia_scr_read(&ap->link, SCR_ERROR, &serror); | |
599 | if (serror == 0x1000500) { | |
600 | ata_port_warn(ap, "Incompatible drive: enabling workaround. This slows down transfer rate to ~60 MB/s"); | |
601 | svia_wd_fix(pdev); | |
602 | hpriv->wd_workaround = true; | |
603 | ap->link.eh_context.i.flags |= ATA_EHI_QUIET; | |
604 | } | |
605 | } | |
606 | ||
607 | ata_sff_error_handler(ap); | |
608 | } | |
609 | ||
610 | static void svia_configure(struct pci_dev *pdev, int board_id, | |
611 | struct svia_priv *hpriv) | |
612 | { | |
613 | u8 tmp8; | |
614 | ||
615 | pci_read_config_byte(pdev, PCI_INTERRUPT_LINE, &tmp8); | |
616 | dev_info(&pdev->dev, "routed to hard irq line %d\n", | |
617 | (int) (tmp8 & 0xf0) == 0xf0 ? 0 : tmp8 & 0x0f); | |
618 | ||
619 | /* make sure SATA channels are enabled */ | |
620 | pci_read_config_byte(pdev, SATA_CHAN_ENAB, &tmp8); | |
621 | if ((tmp8 & ALL_PORTS) != ALL_PORTS) { | |
622 | dev_dbg(&pdev->dev, "enabling SATA channels (0x%x)\n", | |
623 | (int)tmp8); | |
624 | tmp8 |= ALL_PORTS; | |
625 | pci_write_config_byte(pdev, SATA_CHAN_ENAB, tmp8); | |
626 | } | |
627 | ||
628 | /* make sure interrupts for each channel sent to us */ | |
629 | pci_read_config_byte(pdev, SATA_INT_GATE, &tmp8); | |
630 | if ((tmp8 & ALL_PORTS) != ALL_PORTS) { | |
631 | dev_dbg(&pdev->dev, "enabling SATA channel interrupts (0x%x)\n", | |
632 | (int) tmp8); | |
633 | tmp8 |= ALL_PORTS; | |
634 | pci_write_config_byte(pdev, SATA_INT_GATE, tmp8); | |
635 | } | |
636 | ||
637 | /* make sure native mode is enabled */ | |
638 | pci_read_config_byte(pdev, SATA_NATIVE_MODE, &tmp8); | |
639 | if ((tmp8 & NATIVE_MODE_ALL) != NATIVE_MODE_ALL) { | |
640 | dev_dbg(&pdev->dev, | |
641 | "enabling SATA channel native mode (0x%x)\n", | |
642 | (int) tmp8); | |
643 | tmp8 |= NATIVE_MODE_ALL; | |
644 | pci_write_config_byte(pdev, SATA_NATIVE_MODE, tmp8); | |
645 | } | |
646 | ||
647 | /* enable IRQ on hotplug */ | |
648 | pci_read_config_byte(pdev, SVIA_MISC_3, &tmp8); | |
649 | if ((tmp8 & SATA_HOTPLUG) != SATA_HOTPLUG) { | |
650 | dev_dbg(&pdev->dev, | |
651 | "enabling SATA hotplug (0x%x)\n", | |
652 | (int) tmp8); | |
653 | tmp8 |= SATA_HOTPLUG; | |
654 | pci_write_config_byte(pdev, SVIA_MISC_3, tmp8); | |
655 | } | |
656 | ||
657 | /* | |
658 | * vt6420/1 has problems talking to some drives. The following | |
659 | * is the fix from Joseph Chan <JosephChan@via.com.tw>. | |
660 | * | |
661 | * When host issues HOLD, device may send up to 20DW of data | |
662 | * before acknowledging it with HOLDA and the host should be | |
663 | * able to buffer them in FIFO. Unfortunately, some WD drives | |
664 | * send up to 40DW before acknowledging HOLD and, in the | |
665 | * default configuration, this ends up overflowing vt6421's | |
666 | * FIFO, making the controller abort the transaction with | |
667 | * R_ERR. | |
668 | * | |
669 | * Rx52[2] is the internal 128DW FIFO Flow control watermark | |
670 | * adjusting mechanism enable bit and the default value 0 | |
671 | * means host will issue HOLD to device when the left FIFO | |
672 | * size goes below 32DW. Setting it to 1 makes the watermark | |
673 | * 64DW. | |
674 | * | |
675 | * https://bugzilla.kernel.org/show_bug.cgi?id=15173 | |
676 | * http://article.gmane.org/gmane.linux.ide/46352 | |
677 | * http://thread.gmane.org/gmane.linux.kernel/1062139 | |
678 | * | |
679 | * As the fix slows down data transfer, apply it only if the error | |
680 | * actually appears - see vt6421_error_handler() | |
681 | * Apply the fix always on vt6420 as we don't know if SCR_ERROR can be | |
682 | * read safely. | |
683 | */ | |
684 | if (board_id == vt6420) { | |
685 | svia_wd_fix(pdev); | |
686 | hpriv->wd_workaround = true; | |
687 | } | |
688 | } | |
689 | ||
690 | static int svia_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |
691 | { | |
692 | unsigned int i; | |
693 | int rc; | |
694 | struct ata_host *host = NULL; | |
695 | int board_id = (int) ent->driver_data; | |
696 | const unsigned *bar_sizes; | |
697 | struct svia_priv *hpriv; | |
698 | ||
699 | ata_print_version_once(&pdev->dev, DRV_VERSION); | |
700 | ||
701 | rc = pcim_enable_device(pdev); | |
702 | if (rc) | |
703 | return rc; | |
704 | ||
705 | if (board_id == vt6421) | |
706 | bar_sizes = &vt6421_bar_sizes[0]; | |
707 | else | |
708 | bar_sizes = &svia_bar_sizes[0]; | |
709 | ||
710 | for (i = 0; i < ARRAY_SIZE(svia_bar_sizes); i++) | |
711 | if ((pci_resource_start(pdev, i) == 0) || | |
712 | (pci_resource_len(pdev, i) < bar_sizes[i])) { | |
713 | dev_err(&pdev->dev, | |
714 | "invalid PCI BAR %u (sz 0x%llx, val 0x%llx)\n", | |
715 | i, | |
716 | (unsigned long long)pci_resource_start(pdev, i), | |
717 | (unsigned long long)pci_resource_len(pdev, i)); | |
718 | return -ENODEV; | |
719 | } | |
720 | ||
721 | switch (board_id) { | |
722 | case vt6420: | |
723 | rc = vt6420_prepare_host(pdev, &host); | |
724 | break; | |
725 | case vt6421: | |
726 | rc = vt6421_prepare_host(pdev, &host); | |
727 | break; | |
728 | case vt8251: | |
729 | rc = vt8251_prepare_host(pdev, &host); | |
730 | break; | |
731 | default: | |
732 | rc = -EINVAL; | |
733 | } | |
734 | if (rc) | |
735 | return rc; | |
736 | ||
737 | hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL); | |
738 | if (!hpriv) | |
739 | return -ENOMEM; | |
740 | host->private_data = hpriv; | |
741 | ||
742 | svia_configure(pdev, board_id, hpriv); | |
743 | ||
744 | pci_set_master(pdev); | |
745 | if (board_id == vt6421) | |
746 | return ata_host_activate(host, pdev->irq, vt6421_interrupt, | |
747 | IRQF_SHARED, &svia_sht); | |
748 | else | |
749 | return ata_host_activate(host, pdev->irq, ata_bmdma_interrupt, | |
750 | IRQF_SHARED, &svia_sht); | |
751 | } | |
752 | ||
753 | #ifdef CONFIG_PM_SLEEP | |
754 | static int svia_pci_device_resume(struct pci_dev *pdev) | |
755 | { | |
756 | struct ata_host *host = pci_get_drvdata(pdev); | |
757 | struct svia_priv *hpriv = host->private_data; | |
758 | int rc; | |
759 | ||
760 | rc = ata_pci_device_do_resume(pdev); | |
761 | if (rc) | |
762 | return rc; | |
763 | ||
764 | if (hpriv->wd_workaround) | |
765 | svia_wd_fix(pdev); | |
766 | ata_host_resume(host); | |
767 | ||
768 | return 0; | |
769 | } | |
770 | #endif | |
771 | ||
772 | module_pci_driver(svia_pci_driver); |