]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blame - drivers/ata/sata_nv.c
[libata] sata_mv: Fix 50xx irq mask
[mirror_ubuntu-artful-kernel.git] / drivers / ata / sata_nv.c
CommitLineData
1da177e4
LT
1/*
2 * sata_nv.c - NVIDIA nForce SATA
3 *
4 * Copyright 2004 NVIDIA Corp. All rights reserved.
5 * Copyright 2004 Andrew Chew
6 *
aa7e16d6
JG
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License as published by
10 * the Free Software Foundation; either version 2, or (at your option)
11 * any later version.
12 *
13 * This program is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 * GNU General Public License for more details.
17 *
18 * You should have received a copy of the GNU General Public License
19 * along with this program; see the file COPYING. If not, write to
20 * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
1da177e4 21 *
af36d7f0
JG
22 *
23 * libata documentation is available via 'make {ps|pdf}docs',
24 * as Documentation/DocBook/libata.*
25 *
26 * No hardware documentation available outside of NVIDIA.
27 * This driver programs the NVIDIA SATA controller in a similar
28 * fashion as with other PCI IDE BMDMA controllers, with a few
29 * NV-specific details such as register offsets, SATA phy location,
30 * hotplug info, etc.
31 *
fbbb262d
RH
32 * CK804/MCP04 controllers support an alternate programming interface
33 * similar to the ADMA specification (with some modifications).
34 * This allows the use of NCQ. Non-DMA-mapped ATA commands are still
35 * sent through the legacy interface.
36 *
1da177e4
LT
37 */
38
1da177e4
LT
39#include <linux/kernel.h>
40#include <linux/module.h>
41#include <linux/pci.h>
42#include <linux/init.h>
43#include <linux/blkdev.h>
44#include <linux/delay.h>
45#include <linux/interrupt.h>
a9524a76 46#include <linux/device.h>
1da177e4 47#include <scsi/scsi_host.h>
fbbb262d 48#include <scsi/scsi_device.h>
1da177e4
LT
49#include <linux/libata.h>
50
51#define DRV_NAME "sata_nv"
cdf56bcf 52#define DRV_VERSION "3.3"
fbbb262d
RH
53
54#define NV_ADMA_DMA_BOUNDARY 0xffffffffUL
1da177e4 55
10ad05df 56enum {
0d5ff566
TH
57 NV_MMIO_BAR = 5,
58
10ad05df
JG
59 NV_PORTS = 2,
60 NV_PIO_MASK = 0x1f,
61 NV_MWDMA_MASK = 0x07,
62 NV_UDMA_MASK = 0x7f,
63 NV_PORT0_SCR_REG_OFFSET = 0x00,
64 NV_PORT1_SCR_REG_OFFSET = 0x40,
1da177e4 65
27e4b274 66 /* INT_STATUS/ENABLE */
10ad05df 67 NV_INT_STATUS = 0x10,
10ad05df 68 NV_INT_ENABLE = 0x11,
27e4b274 69 NV_INT_STATUS_CK804 = 0x440,
10ad05df 70 NV_INT_ENABLE_CK804 = 0x441,
1da177e4 71
27e4b274
TH
72 /* INT_STATUS/ENABLE bits */
73 NV_INT_DEV = 0x01,
74 NV_INT_PM = 0x02,
75 NV_INT_ADDED = 0x04,
76 NV_INT_REMOVED = 0x08,
77
78 NV_INT_PORT_SHIFT = 4, /* each port occupies 4 bits */
79
39f87582 80 NV_INT_ALL = 0x0f,
5a44efff
TH
81 NV_INT_MASK = NV_INT_DEV |
82 NV_INT_ADDED | NV_INT_REMOVED,
39f87582 83
27e4b274 84 /* INT_CONFIG */
10ad05df
JG
85 NV_INT_CONFIG = 0x12,
86 NV_INT_CONFIG_METHD = 0x01, // 0 = INT, 1 = SMI
1da177e4 87
10ad05df
JG
88 // For PCI config register 20
89 NV_MCP_SATA_CFG_20 = 0x50,
90 NV_MCP_SATA_CFG_20_SATA_SPACE_EN = 0x04,
fbbb262d
RH
91 NV_MCP_SATA_CFG_20_PORT0_EN = (1 << 17),
92 NV_MCP_SATA_CFG_20_PORT1_EN = (1 << 16),
93 NV_MCP_SATA_CFG_20_PORT0_PWB_EN = (1 << 14),
94 NV_MCP_SATA_CFG_20_PORT1_PWB_EN = (1 << 12),
95
96 NV_ADMA_MAX_CPBS = 32,
97 NV_ADMA_CPB_SZ = 128,
98 NV_ADMA_APRD_SZ = 16,
99 NV_ADMA_SGTBL_LEN = (1024 - NV_ADMA_CPB_SZ) /
100 NV_ADMA_APRD_SZ,
101 NV_ADMA_SGTBL_TOTAL_LEN = NV_ADMA_SGTBL_LEN + 5,
102 NV_ADMA_SGTBL_SZ = NV_ADMA_SGTBL_LEN * NV_ADMA_APRD_SZ,
103 NV_ADMA_PORT_PRIV_DMA_SZ = NV_ADMA_MAX_CPBS *
104 (NV_ADMA_CPB_SZ + NV_ADMA_SGTBL_SZ),
105
106 /* BAR5 offset to ADMA general registers */
107 NV_ADMA_GEN = 0x400,
108 NV_ADMA_GEN_CTL = 0x00,
109 NV_ADMA_NOTIFIER_CLEAR = 0x30,
110
111 /* BAR5 offset to ADMA ports */
112 NV_ADMA_PORT = 0x480,
113
114 /* size of ADMA port register space */
115 NV_ADMA_PORT_SIZE = 0x100,
116
117 /* ADMA port registers */
118 NV_ADMA_CTL = 0x40,
119 NV_ADMA_CPB_COUNT = 0x42,
120 NV_ADMA_NEXT_CPB_IDX = 0x43,
121 NV_ADMA_STAT = 0x44,
122 NV_ADMA_CPB_BASE_LOW = 0x48,
123 NV_ADMA_CPB_BASE_HIGH = 0x4C,
124 NV_ADMA_APPEND = 0x50,
125 NV_ADMA_NOTIFIER = 0x68,
126 NV_ADMA_NOTIFIER_ERROR = 0x6C,
127
128 /* NV_ADMA_CTL register bits */
129 NV_ADMA_CTL_HOTPLUG_IEN = (1 << 0),
130 NV_ADMA_CTL_CHANNEL_RESET = (1 << 5),
131 NV_ADMA_CTL_GO = (1 << 7),
132 NV_ADMA_CTL_AIEN = (1 << 8),
133 NV_ADMA_CTL_READ_NON_COHERENT = (1 << 11),
134 NV_ADMA_CTL_WRITE_NON_COHERENT = (1 << 12),
135
136 /* CPB response flag bits */
137 NV_CPB_RESP_DONE = (1 << 0),
138 NV_CPB_RESP_ATA_ERR = (1 << 3),
139 NV_CPB_RESP_CMD_ERR = (1 << 4),
140 NV_CPB_RESP_CPB_ERR = (1 << 7),
141
142 /* CPB control flag bits */
143 NV_CPB_CTL_CPB_VALID = (1 << 0),
144 NV_CPB_CTL_QUEUE = (1 << 1),
145 NV_CPB_CTL_APRD_VALID = (1 << 2),
146 NV_CPB_CTL_IEN = (1 << 3),
147 NV_CPB_CTL_FPDMA = (1 << 4),
148
149 /* APRD flags */
150 NV_APRD_WRITE = (1 << 1),
151 NV_APRD_END = (1 << 2),
152 NV_APRD_CONT = (1 << 3),
153
154 /* NV_ADMA_STAT flags */
155 NV_ADMA_STAT_TIMEOUT = (1 << 0),
156 NV_ADMA_STAT_HOTUNPLUG = (1 << 1),
157 NV_ADMA_STAT_HOTPLUG = (1 << 2),
158 NV_ADMA_STAT_CPBERR = (1 << 4),
159 NV_ADMA_STAT_SERROR = (1 << 5),
160 NV_ADMA_STAT_CMD_COMPLETE = (1 << 6),
161 NV_ADMA_STAT_IDLE = (1 << 8),
162 NV_ADMA_STAT_LEGACY = (1 << 9),
163 NV_ADMA_STAT_STOPPED = (1 << 10),
164 NV_ADMA_STAT_DONE = (1 << 12),
165 NV_ADMA_STAT_ERR = NV_ADMA_STAT_CPBERR |
166 NV_ADMA_STAT_TIMEOUT,
167
168 /* port flags */
169 NV_ADMA_PORT_REGISTER_MODE = (1 << 0),
2dec7555 170 NV_ADMA_ATAPI_SETUP_COMPLETE = (1 << 1),
fbbb262d
RH
171
172};
173
174/* ADMA Physical Region Descriptor - one SG segment */
175struct nv_adma_prd {
176 __le64 addr;
177 __le32 len;
178 u8 flags;
179 u8 packet_len;
180 __le16 reserved;
181};
182
183enum nv_adma_regbits {
184 CMDEND = (1 << 15), /* end of command list */
185 WNB = (1 << 14), /* wait-not-BSY */
186 IGN = (1 << 13), /* ignore this entry */
187 CS1n = (1 << (4 + 8)), /* std. PATA signals follow... */
188 DA2 = (1 << (2 + 8)),
189 DA1 = (1 << (1 + 8)),
190 DA0 = (1 << (0 + 8)),
191};
192
193/* ADMA Command Parameter Block
194 The first 5 SG segments are stored inside the Command Parameter Block itself.
195 If there are more than 5 segments the remainder are stored in a separate
196 memory area indicated by next_aprd. */
197struct nv_adma_cpb {
198 u8 resp_flags; /* 0 */
199 u8 reserved1; /* 1 */
200 u8 ctl_flags; /* 2 */
201 /* len is length of taskfile in 64 bit words */
202 u8 len; /* 3 */
203 u8 tag; /* 4 */
204 u8 next_cpb_idx; /* 5 */
205 __le16 reserved2; /* 6-7 */
206 __le16 tf[12]; /* 8-31 */
207 struct nv_adma_prd aprd[5]; /* 32-111 */
208 __le64 next_aprd; /* 112-119 */
209 __le64 reserved3; /* 120-127 */
10ad05df 210};
1da177e4 211
fbbb262d
RH
212
213struct nv_adma_port_priv {
214 struct nv_adma_cpb *cpb;
215 dma_addr_t cpb_dma;
216 struct nv_adma_prd *aprd;
217 dma_addr_t aprd_dma;
cdf56bcf
RH
218 void __iomem * ctl_block;
219 void __iomem * gen_block;
220 void __iomem * notifier_clear_block;
fbbb262d 221 u8 flags;
5e5c74a5 222 int last_issue_ncq;
fbbb262d
RH
223};
224
cdf56bcf
RH
225struct nv_host_priv {
226 unsigned long type;
227};
228
fbbb262d
RH
229#define NV_ADMA_CHECK_INTR(GCTL, PORT) ((GCTL) & ( 1 << (19 + (12 * (PORT)))))
230
1da177e4 231static int nv_init_one (struct pci_dev *pdev, const struct pci_device_id *ent);
cdf56bcf
RH
232static void nv_remove_one (struct pci_dev *pdev);
233static int nv_pci_device_resume(struct pci_dev *pdev);
cca3974e 234static void nv_ck804_host_stop(struct ata_host *host);
7d12e780
DH
235static irqreturn_t nv_generic_interrupt(int irq, void *dev_instance);
236static irqreturn_t nv_nf2_interrupt(int irq, void *dev_instance);
237static irqreturn_t nv_ck804_interrupt(int irq, void *dev_instance);
1da177e4
LT
238static u32 nv_scr_read (struct ata_port *ap, unsigned int sc_reg);
239static void nv_scr_write (struct ata_port *ap, unsigned int sc_reg, u32 val);
1da177e4 240
39f87582
TH
241static void nv_nf2_freeze(struct ata_port *ap);
242static void nv_nf2_thaw(struct ata_port *ap);
243static void nv_ck804_freeze(struct ata_port *ap);
244static void nv_ck804_thaw(struct ata_port *ap);
245static void nv_error_handler(struct ata_port *ap);
fbbb262d 246static int nv_adma_slave_config(struct scsi_device *sdev);
2dec7555 247static int nv_adma_check_atapi_dma(struct ata_queued_cmd *qc);
fbbb262d
RH
248static void nv_adma_qc_prep(struct ata_queued_cmd *qc);
249static unsigned int nv_adma_qc_issue(struct ata_queued_cmd *qc);
250static irqreturn_t nv_adma_interrupt(int irq, void *dev_instance);
251static void nv_adma_irq_clear(struct ata_port *ap);
252static int nv_adma_port_start(struct ata_port *ap);
253static void nv_adma_port_stop(struct ata_port *ap);
cdf56bcf
RH
254static int nv_adma_port_suspend(struct ata_port *ap, pm_message_t mesg);
255static int nv_adma_port_resume(struct ata_port *ap);
fbbb262d
RH
256static void nv_adma_error_handler(struct ata_port *ap);
257static void nv_adma_host_stop(struct ata_host *host);
f5ecac2d 258static void nv_adma_post_internal_cmd(struct ata_queued_cmd *qc);
39f87582 259
1da177e4
LT
260enum nv_host_type
261{
262 GENERIC,
263 NFORCE2,
27e4b274 264 NFORCE3 = NFORCE2, /* NF2 == NF3 as far as sata_nv is concerned */
fbbb262d
RH
265 CK804,
266 ADMA
1da177e4
LT
267};
268
3b7d697d 269static const struct pci_device_id nv_pci_tbl[] = {
54bb3a94
JG
270 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE2S_SATA), NFORCE2 },
271 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE3S_SATA), NFORCE3 },
272 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE3S_SATA2), NFORCE3 },
273 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_CK804_SATA), CK804 },
274 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_CK804_SATA2), CK804 },
275 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP04_SATA), CK804 },
276 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP04_SATA2), CK804 },
277 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP51_SATA), GENERIC },
278 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP51_SATA2), GENERIC },
279 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP55_SATA), GENERIC },
280 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP55_SATA2), GENERIC },
281 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP61_SATA), GENERIC },
282 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP61_SATA2), GENERIC },
283 { PCI_VDEVICE(NVIDIA, PCI_DEVICE_ID_NVIDIA_NFORCE_MCP61_SATA3), GENERIC },
1da177e4
LT
284 { PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID,
285 PCI_ANY_ID, PCI_ANY_ID,
286 PCI_CLASS_STORAGE_IDE<<8, 0xffff00, GENERIC },
541134cf
DD
287 { PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID,
288 PCI_ANY_ID, PCI_ANY_ID,
289 PCI_CLASS_STORAGE_RAID<<8, 0xffff00, GENERIC },
2d2744fc
JG
290
291 { } /* terminate list */
1da177e4
LT
292};
293
1da177e4
LT
294static struct pci_driver nv_pci_driver = {
295 .name = DRV_NAME,
296 .id_table = nv_pci_tbl,
297 .probe = nv_init_one,
cdf56bcf
RH
298 .suspend = ata_pci_device_suspend,
299 .resume = nv_pci_device_resume,
300 .remove = nv_remove_one,
1da177e4
LT
301};
302
193515d5 303static struct scsi_host_template nv_sht = {
1da177e4
LT
304 .module = THIS_MODULE,
305 .name = DRV_NAME,
306 .ioctl = ata_scsi_ioctl,
307 .queuecommand = ata_scsi_queuecmd,
1da177e4
LT
308 .can_queue = ATA_DEF_QUEUE,
309 .this_id = ATA_SHT_THIS_ID,
310 .sg_tablesize = LIBATA_MAX_PRD,
1da177e4
LT
311 .cmd_per_lun = ATA_SHT_CMD_PER_LUN,
312 .emulated = ATA_SHT_EMULATED,
313 .use_clustering = ATA_SHT_USE_CLUSTERING,
314 .proc_name = DRV_NAME,
315 .dma_boundary = ATA_DMA_BOUNDARY,
316 .slave_configure = ata_scsi_slave_config,
ccf68c34 317 .slave_destroy = ata_scsi_slave_destroy,
1da177e4 318 .bios_param = ata_std_bios_param,
cdf56bcf
RH
319 .suspend = ata_scsi_device_suspend,
320 .resume = ata_scsi_device_resume,
1da177e4
LT
321};
322
fbbb262d
RH
323static struct scsi_host_template nv_adma_sht = {
324 .module = THIS_MODULE,
325 .name = DRV_NAME,
326 .ioctl = ata_scsi_ioctl,
327 .queuecommand = ata_scsi_queuecmd,
328 .can_queue = NV_ADMA_MAX_CPBS,
329 .this_id = ATA_SHT_THIS_ID,
330 .sg_tablesize = NV_ADMA_SGTBL_TOTAL_LEN,
fbbb262d
RH
331 .cmd_per_lun = ATA_SHT_CMD_PER_LUN,
332 .emulated = ATA_SHT_EMULATED,
333 .use_clustering = ATA_SHT_USE_CLUSTERING,
334 .proc_name = DRV_NAME,
335 .dma_boundary = NV_ADMA_DMA_BOUNDARY,
336 .slave_configure = nv_adma_slave_config,
337 .slave_destroy = ata_scsi_slave_destroy,
338 .bios_param = ata_std_bios_param,
cdf56bcf
RH
339 .suspend = ata_scsi_device_suspend,
340 .resume = ata_scsi_device_resume,
fbbb262d
RH
341};
342
ada364e8 343static const struct ata_port_operations nv_generic_ops = {
1da177e4
LT
344 .port_disable = ata_port_disable,
345 .tf_load = ata_tf_load,
346 .tf_read = ata_tf_read,
347 .exec_command = ata_exec_command,
348 .check_status = ata_check_status,
349 .dev_select = ata_std_dev_select,
1da177e4
LT
350 .bmdma_setup = ata_bmdma_setup,
351 .bmdma_start = ata_bmdma_start,
352 .bmdma_stop = ata_bmdma_stop,
353 .bmdma_status = ata_bmdma_status,
354 .qc_prep = ata_qc_prep,
355 .qc_issue = ata_qc_issue_prot,
39f87582
TH
356 .freeze = ata_bmdma_freeze,
357 .thaw = ata_bmdma_thaw,
358 .error_handler = nv_error_handler,
359 .post_internal_cmd = ata_bmdma_post_internal_cmd,
0d5ff566 360 .data_xfer = ata_data_xfer,
ada364e8 361 .irq_handler = nv_generic_interrupt,
1da177e4 362 .irq_clear = ata_bmdma_irq_clear,
246ce3b6
AI
363 .irq_on = ata_irq_on,
364 .irq_ack = ata_irq_ack,
1da177e4
LT
365 .scr_read = nv_scr_read,
366 .scr_write = nv_scr_write,
367 .port_start = ata_port_start,
1da177e4
LT
368};
369
ada364e8
TH
370static const struct ata_port_operations nv_nf2_ops = {
371 .port_disable = ata_port_disable,
372 .tf_load = ata_tf_load,
373 .tf_read = ata_tf_read,
374 .exec_command = ata_exec_command,
375 .check_status = ata_check_status,
376 .dev_select = ata_std_dev_select,
ada364e8
TH
377 .bmdma_setup = ata_bmdma_setup,
378 .bmdma_start = ata_bmdma_start,
379 .bmdma_stop = ata_bmdma_stop,
380 .bmdma_status = ata_bmdma_status,
381 .qc_prep = ata_qc_prep,
382 .qc_issue = ata_qc_issue_prot,
39f87582
TH
383 .freeze = nv_nf2_freeze,
384 .thaw = nv_nf2_thaw,
385 .error_handler = nv_error_handler,
386 .post_internal_cmd = ata_bmdma_post_internal_cmd,
0d5ff566 387 .data_xfer = ata_data_xfer,
ada364e8
TH
388 .irq_handler = nv_nf2_interrupt,
389 .irq_clear = ata_bmdma_irq_clear,
246ce3b6
AI
390 .irq_on = ata_irq_on,
391 .irq_ack = ata_irq_ack,
ada364e8
TH
392 .scr_read = nv_scr_read,
393 .scr_write = nv_scr_write,
394 .port_start = ata_port_start,
ada364e8
TH
395};
396
397static const struct ata_port_operations nv_ck804_ops = {
398 .port_disable = ata_port_disable,
399 .tf_load = ata_tf_load,
400 .tf_read = ata_tf_read,
401 .exec_command = ata_exec_command,
402 .check_status = ata_check_status,
403 .dev_select = ata_std_dev_select,
ada364e8
TH
404 .bmdma_setup = ata_bmdma_setup,
405 .bmdma_start = ata_bmdma_start,
406 .bmdma_stop = ata_bmdma_stop,
407 .bmdma_status = ata_bmdma_status,
408 .qc_prep = ata_qc_prep,
409 .qc_issue = ata_qc_issue_prot,
39f87582
TH
410 .freeze = nv_ck804_freeze,
411 .thaw = nv_ck804_thaw,
412 .error_handler = nv_error_handler,
413 .post_internal_cmd = ata_bmdma_post_internal_cmd,
0d5ff566 414 .data_xfer = ata_data_xfer,
ada364e8
TH
415 .irq_handler = nv_ck804_interrupt,
416 .irq_clear = ata_bmdma_irq_clear,
246ce3b6
AI
417 .irq_on = ata_irq_on,
418 .irq_ack = ata_irq_ack,
ada364e8
TH
419 .scr_read = nv_scr_read,
420 .scr_write = nv_scr_write,
421 .port_start = ata_port_start,
ada364e8
TH
422 .host_stop = nv_ck804_host_stop,
423};
424
fbbb262d
RH
425static const struct ata_port_operations nv_adma_ops = {
426 .port_disable = ata_port_disable,
427 .tf_load = ata_tf_load,
428 .tf_read = ata_tf_read,
2dec7555 429 .check_atapi_dma = nv_adma_check_atapi_dma,
fbbb262d
RH
430 .exec_command = ata_exec_command,
431 .check_status = ata_check_status,
432 .dev_select = ata_std_dev_select,
f5ecac2d
RH
433 .bmdma_setup = ata_bmdma_setup,
434 .bmdma_start = ata_bmdma_start,
435 .bmdma_stop = ata_bmdma_stop,
436 .bmdma_status = ata_bmdma_status,
fbbb262d
RH
437 .qc_prep = nv_adma_qc_prep,
438 .qc_issue = nv_adma_qc_issue,
439 .freeze = nv_ck804_freeze,
440 .thaw = nv_ck804_thaw,
441 .error_handler = nv_adma_error_handler,
f5ecac2d 442 .post_internal_cmd = nv_adma_post_internal_cmd,
0d5ff566 443 .data_xfer = ata_data_xfer,
fbbb262d
RH
444 .irq_handler = nv_adma_interrupt,
445 .irq_clear = nv_adma_irq_clear,
246ce3b6
AI
446 .irq_on = ata_irq_on,
447 .irq_ack = ata_irq_ack,
fbbb262d
RH
448 .scr_read = nv_scr_read,
449 .scr_write = nv_scr_write,
450 .port_start = nv_adma_port_start,
451 .port_stop = nv_adma_port_stop,
cdf56bcf
RH
452 .port_suspend = nv_adma_port_suspend,
453 .port_resume = nv_adma_port_resume,
fbbb262d
RH
454 .host_stop = nv_adma_host_stop,
455};
456
ada364e8
TH
457static struct ata_port_info nv_port_info[] = {
458 /* generic */
459 {
460 .sht = &nv_sht,
722420fe
TH
461 .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY |
462 ATA_FLAG_HRST_TO_RESUME,
ada364e8
TH
463 .pio_mask = NV_PIO_MASK,
464 .mwdma_mask = NV_MWDMA_MASK,
465 .udma_mask = NV_UDMA_MASK,
466 .port_ops = &nv_generic_ops,
467 },
468 /* nforce2/3 */
469 {
470 .sht = &nv_sht,
722420fe
TH
471 .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY |
472 ATA_FLAG_HRST_TO_RESUME,
ada364e8
TH
473 .pio_mask = NV_PIO_MASK,
474 .mwdma_mask = NV_MWDMA_MASK,
475 .udma_mask = NV_UDMA_MASK,
476 .port_ops = &nv_nf2_ops,
477 },
478 /* ck804 */
479 {
480 .sht = &nv_sht,
722420fe
TH
481 .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY |
482 ATA_FLAG_HRST_TO_RESUME,
ada364e8
TH
483 .pio_mask = NV_PIO_MASK,
484 .mwdma_mask = NV_MWDMA_MASK,
485 .udma_mask = NV_UDMA_MASK,
486 .port_ops = &nv_ck804_ops,
487 },
fbbb262d
RH
488 /* ADMA */
489 {
490 .sht = &nv_adma_sht,
491 .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY |
cdf56bcf 492 ATA_FLAG_HRST_TO_RESUME |
fbbb262d
RH
493 ATA_FLAG_MMIO | ATA_FLAG_NCQ,
494 .pio_mask = NV_PIO_MASK,
495 .mwdma_mask = NV_MWDMA_MASK,
496 .udma_mask = NV_UDMA_MASK,
497 .port_ops = &nv_adma_ops,
498 },
1da177e4
LT
499};
500
501MODULE_AUTHOR("NVIDIA");
502MODULE_DESCRIPTION("low-level driver for NVIDIA nForce SATA controller");
503MODULE_LICENSE("GPL");
504MODULE_DEVICE_TABLE(pci, nv_pci_tbl);
505MODULE_VERSION(DRV_VERSION);
506
fbbb262d
RH
507static int adma_enabled = 1;
508
2dec7555
RH
509static void nv_adma_register_mode(struct ata_port *ap)
510{
2dec7555 511 struct nv_adma_port_priv *pp = ap->private_data;
cdf56bcf 512 void __iomem *mmio = pp->ctl_block;
a2cfe81a
RH
513 u16 tmp, status;
514 int count = 0;
2dec7555
RH
515
516 if (pp->flags & NV_ADMA_PORT_REGISTER_MODE)
517 return;
518
a2cfe81a
RH
519 status = readw(mmio + NV_ADMA_STAT);
520 while(!(status & NV_ADMA_STAT_IDLE) && count < 20) {
521 ndelay(50);
522 status = readw(mmio + NV_ADMA_STAT);
523 count++;
524 }
525 if(count == 20)
526 ata_port_printk(ap, KERN_WARNING,
527 "timeout waiting for ADMA IDLE, stat=0x%hx\n",
528 status);
529
2dec7555
RH
530 tmp = readw(mmio + NV_ADMA_CTL);
531 writew(tmp & ~NV_ADMA_CTL_GO, mmio + NV_ADMA_CTL);
532
a2cfe81a
RH
533 count = 0;
534 status = readw(mmio + NV_ADMA_STAT);
535 while(!(status & NV_ADMA_STAT_LEGACY) && count < 20) {
536 ndelay(50);
537 status = readw(mmio + NV_ADMA_STAT);
538 count++;
539 }
540 if(count == 20)
541 ata_port_printk(ap, KERN_WARNING,
542 "timeout waiting for ADMA LEGACY, stat=0x%hx\n",
543 status);
544
2dec7555
RH
545 pp->flags |= NV_ADMA_PORT_REGISTER_MODE;
546}
547
548static void nv_adma_mode(struct ata_port *ap)
549{
2dec7555 550 struct nv_adma_port_priv *pp = ap->private_data;
cdf56bcf 551 void __iomem *mmio = pp->ctl_block;
a2cfe81a
RH
552 u16 tmp, status;
553 int count = 0;
2dec7555
RH
554
555 if (!(pp->flags & NV_ADMA_PORT_REGISTER_MODE))
556 return;
f20b16ff 557
2dec7555
RH
558 WARN_ON(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE);
559
560 tmp = readw(mmio + NV_ADMA_CTL);
561 writew(tmp | NV_ADMA_CTL_GO, mmio + NV_ADMA_CTL);
562
a2cfe81a
RH
563 status = readw(mmio + NV_ADMA_STAT);
564 while(((status & NV_ADMA_STAT_LEGACY) ||
565 !(status & NV_ADMA_STAT_IDLE)) && count < 20) {
566 ndelay(50);
567 status = readw(mmio + NV_ADMA_STAT);
568 count++;
569 }
570 if(count == 20)
571 ata_port_printk(ap, KERN_WARNING,
572 "timeout waiting for ADMA LEGACY clear and IDLE, stat=0x%hx\n",
573 status);
574
2dec7555
RH
575 pp->flags &= ~NV_ADMA_PORT_REGISTER_MODE;
576}
577
fbbb262d
RH
578static int nv_adma_slave_config(struct scsi_device *sdev)
579{
580 struct ata_port *ap = ata_shost_to_port(sdev->host);
2dec7555
RH
581 struct nv_adma_port_priv *pp = ap->private_data;
582 struct pci_dev *pdev = to_pci_dev(ap->host->dev);
fbbb262d
RH
583 u64 bounce_limit;
584 unsigned long segment_boundary;
585 unsigned short sg_tablesize;
586 int rc;
2dec7555
RH
587 int adma_enable;
588 u32 current_reg, new_reg, config_mask;
fbbb262d
RH
589
590 rc = ata_scsi_slave_config(sdev);
591
592 if (sdev->id >= ATA_MAX_DEVICES || sdev->channel || sdev->lun)
593 /* Not a proper libata device, ignore */
594 return rc;
595
596 if (ap->device[sdev->id].class == ATA_DEV_ATAPI) {
597 /*
598 * NVIDIA reports that ADMA mode does not support ATAPI commands.
599 * Therefore ATAPI commands are sent through the legacy interface.
600 * However, the legacy interface only supports 32-bit DMA.
601 * Restrict DMA parameters as required by the legacy interface
602 * when an ATAPI device is connected.
603 */
604 bounce_limit = ATA_DMA_MASK;
605 segment_boundary = ATA_DMA_BOUNDARY;
606 /* Subtract 1 since an extra entry may be needed for padding, see
607 libata-scsi.c */
608 sg_tablesize = LIBATA_MAX_PRD - 1;
f20b16ff 609
2dec7555
RH
610 /* Since the legacy DMA engine is in use, we need to disable ADMA
611 on the port. */
612 adma_enable = 0;
613 nv_adma_register_mode(ap);
fbbb262d
RH
614 }
615 else {
616 bounce_limit = *ap->dev->dma_mask;
617 segment_boundary = NV_ADMA_DMA_BOUNDARY;
618 sg_tablesize = NV_ADMA_SGTBL_TOTAL_LEN;
2dec7555 619 adma_enable = 1;
fbbb262d 620 }
f20b16ff 621
2dec7555
RH
622 pci_read_config_dword(pdev, NV_MCP_SATA_CFG_20, &current_reg);
623
624 if(ap->port_no == 1)
625 config_mask = NV_MCP_SATA_CFG_20_PORT1_EN |
626 NV_MCP_SATA_CFG_20_PORT1_PWB_EN;
627 else
628 config_mask = NV_MCP_SATA_CFG_20_PORT0_EN |
629 NV_MCP_SATA_CFG_20_PORT0_PWB_EN;
f20b16ff 630
2dec7555
RH
631 if(adma_enable) {
632 new_reg = current_reg | config_mask;
633 pp->flags &= ~NV_ADMA_ATAPI_SETUP_COMPLETE;
634 }
635 else {
636 new_reg = current_reg & ~config_mask;
637 pp->flags |= NV_ADMA_ATAPI_SETUP_COMPLETE;
638 }
f20b16ff 639
2dec7555
RH
640 if(current_reg != new_reg)
641 pci_write_config_dword(pdev, NV_MCP_SATA_CFG_20, new_reg);
f20b16ff 642
fbbb262d
RH
643 blk_queue_bounce_limit(sdev->request_queue, bounce_limit);
644 blk_queue_segment_boundary(sdev->request_queue, segment_boundary);
645 blk_queue_max_hw_segments(sdev->request_queue, sg_tablesize);
646 ata_port_printk(ap, KERN_INFO,
647 "bounce limit 0x%llX, segment boundary 0x%lX, hw segs %hu\n",
648 (unsigned long long)bounce_limit, segment_boundary, sg_tablesize);
649 return rc;
650}
651
2dec7555
RH
652static int nv_adma_check_atapi_dma(struct ata_queued_cmd *qc)
653{
654 struct nv_adma_port_priv *pp = qc->ap->private_data;
655 return !(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE);
656}
657
658static unsigned int nv_adma_tf_to_cpb(struct ata_taskfile *tf, __le16 *cpb)
fbbb262d
RH
659{
660 unsigned int idx = 0;
661
ac3d6b86
RH
662 if(tf->flags & ATA_TFLAG_ISADDR) {
663 if (tf->flags & ATA_TFLAG_LBA48) {
664 cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->hob_feature | WNB);
665 cpb[idx++] = cpu_to_le16((ATA_REG_NSECT << 8) | tf->hob_nsect);
666 cpb[idx++] = cpu_to_le16((ATA_REG_LBAL << 8) | tf->hob_lbal);
667 cpb[idx++] = cpu_to_le16((ATA_REG_LBAM << 8) | tf->hob_lbam);
668 cpb[idx++] = cpu_to_le16((ATA_REG_LBAH << 8) | tf->hob_lbah);
669 cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->feature);
670 } else
671 cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->feature | WNB);
672
673 cpb[idx++] = cpu_to_le16((ATA_REG_NSECT << 8) | tf->nsect);
674 cpb[idx++] = cpu_to_le16((ATA_REG_LBAL << 8) | tf->lbal);
675 cpb[idx++] = cpu_to_le16((ATA_REG_LBAM << 8) | tf->lbam);
676 cpb[idx++] = cpu_to_le16((ATA_REG_LBAH << 8) | tf->lbah);
fbbb262d 677 }
ac3d6b86
RH
678
679 if(tf->flags & ATA_TFLAG_DEVICE)
680 cpb[idx++] = cpu_to_le16((ATA_REG_DEVICE << 8) | tf->device);
fbbb262d
RH
681
682 cpb[idx++] = cpu_to_le16((ATA_REG_CMD << 8) | tf->command | CMDEND);
ac3d6b86
RH
683
684 while(idx < 12)
685 cpb[idx++] = cpu_to_le16(IGN);
fbbb262d
RH
686
687 return idx;
688}
689
5bd28a4b 690static int nv_adma_check_cpb(struct ata_port *ap, int cpb_num, int force_err)
fbbb262d
RH
691{
692 struct nv_adma_port_priv *pp = ap->private_data;
2dec7555 693 u8 flags = pp->cpb[cpb_num].resp_flags;
fbbb262d
RH
694
695 VPRINTK("CPB %d, flags=0x%x\n", cpb_num, flags);
696
5bd28a4b
RH
697 if (unlikely((force_err ||
698 flags & (NV_CPB_RESP_ATA_ERR |
699 NV_CPB_RESP_CMD_ERR |
700 NV_CPB_RESP_CPB_ERR)))) {
701 struct ata_eh_info *ehi = &ap->eh_info;
702 int freeze = 0;
703
704 ata_ehi_clear_desc(ehi);
705 ata_ehi_push_desc(ehi, "CPB resp_flags 0x%x", flags );
706 if (flags & NV_CPB_RESP_ATA_ERR) {
707 ata_ehi_push_desc(ehi, ": ATA error");
708 ehi->err_mask |= AC_ERR_DEV;
709 } else if (flags & NV_CPB_RESP_CMD_ERR) {
710 ata_ehi_push_desc(ehi, ": CMD error");
711 ehi->err_mask |= AC_ERR_DEV;
712 } else if (flags & NV_CPB_RESP_CPB_ERR) {
713 ata_ehi_push_desc(ehi, ": CPB error");
714 ehi->err_mask |= AC_ERR_SYSTEM;
715 freeze = 1;
716 } else {
717 /* notifier error, but no error in CPB flags? */
718 ehi->err_mask |= AC_ERR_OTHER;
719 freeze = 1;
720 }
721 /* Kill all commands. EH will determine what actually failed. */
722 if (freeze)
723 ata_port_freeze(ap);
724 else
725 ata_port_abort(ap);
726 return 1;
fbbb262d 727 }
5bd28a4b
RH
728
729 if (flags & NV_CPB_RESP_DONE) {
fbbb262d 730 struct ata_queued_cmd *qc = ata_qc_from_tag(ap, cpb_num);
5bd28a4b
RH
731 VPRINTK("CPB flags done, flags=0x%x\n", flags);
732 if (likely(qc)) {
733 /* Grab the ATA port status for non-NCQ commands.
fbbb262d
RH
734 For NCQ commands the current status may have nothing to do with
735 the command just completed. */
5bd28a4b
RH
736 if (qc->tf.protocol != ATA_PROT_NCQ) {
737 u8 ata_status = readb(pp->ctl_block + (ATA_REG_STATUS * 4));
738 qc->err_mask |= ac_err_mask(ata_status);
739 }
fbbb262d
RH
740 DPRINTK("Completing qc from tag %d with err_mask %u\n",cpb_num,
741 qc->err_mask);
742 ata_qc_complete(qc);
2a54cf76
RH
743 } else {
744 struct ata_eh_info *ehi = &ap->eh_info;
745 /* Notifier bits set without a command may indicate the drive
746 is misbehaving. Raise host state machine violation on this
747 condition. */
748 ata_port_printk(ap, KERN_ERR, "notifier for tag %d with no command?\n",
749 cpb_num);
750 ehi->err_mask |= AC_ERR_HSM;
751 ehi->action |= ATA_EH_SOFTRESET;
752 ata_port_freeze(ap);
753 return 1;
fbbb262d
RH
754 }
755 }
5bd28a4b 756 return 0;
fbbb262d
RH
757}
758
2dec7555
RH
759static int nv_host_intr(struct ata_port *ap, u8 irq_stat)
760{
761 struct ata_queued_cmd *qc = ata_qc_from_tag(ap, ap->active_tag);
2dec7555
RH
762
763 /* freeze if hotplugged */
764 if (unlikely(irq_stat & (NV_INT_ADDED | NV_INT_REMOVED))) {
765 ata_port_freeze(ap);
766 return 1;
767 }
768
769 /* bail out if not our interrupt */
770 if (!(irq_stat & NV_INT_DEV))
771 return 0;
772
773 /* DEV interrupt w/ no active qc? */
774 if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) {
775 ata_check_status(ap);
776 return 1;
777 }
778
779 /* handle interrupt */
f740d168 780 return ata_host_intr(ap, qc);
2dec7555
RH
781}
782
fbbb262d
RH
783static irqreturn_t nv_adma_interrupt(int irq, void *dev_instance)
784{
785 struct ata_host *host = dev_instance;
786 int i, handled = 0;
2dec7555 787 u32 notifier_clears[2];
fbbb262d
RH
788
789 spin_lock(&host->lock);
790
791 for (i = 0; i < host->n_ports; i++) {
792 struct ata_port *ap = host->ports[i];
2dec7555 793 notifier_clears[i] = 0;
fbbb262d
RH
794
795 if (ap && !(ap->flags & ATA_FLAG_DISABLED)) {
796 struct nv_adma_port_priv *pp = ap->private_data;
cdf56bcf 797 void __iomem *mmio = pp->ctl_block;
fbbb262d
RH
798 u16 status;
799 u32 gen_ctl;
fbbb262d
RH
800 u32 notifier, notifier_error;
801
802 /* if in ATA register mode, use standard ata interrupt handler */
803 if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) {
0d5ff566 804 u8 irq_stat = readb(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804)
2dec7555 805 >> (NV_INT_PORT_SHIFT * i);
f740d168
RH
806 if(ata_tag_valid(ap->active_tag))
807 /** NV_INT_DEV indication seems unreliable at times
808 at least in ADMA mode. Force it on always when a
809 command is active, to prevent losing interrupts. */
810 irq_stat |= NV_INT_DEV;
2dec7555 811 handled += nv_host_intr(ap, irq_stat);
fbbb262d
RH
812 continue;
813 }
814
815 notifier = readl(mmio + NV_ADMA_NOTIFIER);
816 notifier_error = readl(mmio + NV_ADMA_NOTIFIER_ERROR);
2dec7555 817 notifier_clears[i] = notifier | notifier_error;
fbbb262d 818
cdf56bcf 819 gen_ctl = readl(pp->gen_block + NV_ADMA_GEN_CTL);
fbbb262d 820
fbbb262d
RH
821 if( !NV_ADMA_CHECK_INTR(gen_ctl, ap->port_no) && !notifier &&
822 !notifier_error)
823 /* Nothing to do */
824 continue;
825
826 status = readw(mmio + NV_ADMA_STAT);
827
828 /* Clear status. Ensure the controller sees the clearing before we start
829 looking at any of the CPB statuses, so that any CPB completions after
830 this point in the handler will raise another interrupt. */
831 writew(status, mmio + NV_ADMA_STAT);
832 readw(mmio + NV_ADMA_STAT); /* flush posted write */
833 rmb();
834
5bd28a4b
RH
835 handled++; /* irq handled if we got here */
836
837 /* freeze if hotplugged or controller error */
838 if (unlikely(status & (NV_ADMA_STAT_HOTPLUG |
839 NV_ADMA_STAT_HOTUNPLUG |
5278b50c
RH
840 NV_ADMA_STAT_TIMEOUT |
841 NV_ADMA_STAT_SERROR))) {
5bd28a4b
RH
842 struct ata_eh_info *ehi = &ap->eh_info;
843
844 ata_ehi_clear_desc(ehi);
845 ata_ehi_push_desc(ehi, "ADMA status 0x%08x", status );
846 if (status & NV_ADMA_STAT_TIMEOUT) {
847 ehi->err_mask |= AC_ERR_SYSTEM;
848 ata_ehi_push_desc(ehi, ": timeout");
849 } else if (status & NV_ADMA_STAT_HOTPLUG) {
850 ata_ehi_hotplugged(ehi);
851 ata_ehi_push_desc(ehi, ": hotplug");
852 } else if (status & NV_ADMA_STAT_HOTUNPLUG) {
853 ata_ehi_hotplugged(ehi);
854 ata_ehi_push_desc(ehi, ": hot unplug");
5278b50c
RH
855 } else if (status & NV_ADMA_STAT_SERROR) {
856 /* let libata analyze SError and figure out the cause */
857 ata_ehi_push_desc(ehi, ": SError");
5bd28a4b 858 }
fbbb262d 859 ata_port_freeze(ap);
fbbb262d
RH
860 continue;
861 }
862
5bd28a4b
RH
863 if (status & (NV_ADMA_STAT_DONE |
864 NV_ADMA_STAT_CPBERR)) {
721449bf
RH
865 u32 check_commands = notifier | notifier_error;
866 int pos, error = 0;
fbbb262d 867 /** Check CPBs for completed commands */
721449bf
RH
868 while ((pos = ffs(check_commands)) && !error) {
869 pos--;
870 error = nv_adma_check_cpb(ap, pos,
871 notifier_error & (1 << pos) );
872 check_commands &= ~(1 << pos );
fbbb262d
RH
873 }
874 }
fbbb262d
RH
875 }
876 }
f20b16ff 877
2dec7555
RH
878 if(notifier_clears[0] || notifier_clears[1]) {
879 /* Note: Both notifier clear registers must be written
880 if either is set, even if one is zero, according to NVIDIA. */
cdf56bcf
RH
881 struct nv_adma_port_priv *pp = host->ports[0]->private_data;
882 writel(notifier_clears[0], pp->notifier_clear_block);
883 pp = host->ports[1]->private_data;
884 writel(notifier_clears[1], pp->notifier_clear_block);
2dec7555 885 }
fbbb262d
RH
886
887 spin_unlock(&host->lock);
888
889 return IRQ_RETVAL(handled);
890}
891
892static void nv_adma_irq_clear(struct ata_port *ap)
893{
cdf56bcf
RH
894 struct nv_adma_port_priv *pp = ap->private_data;
895 void __iomem *mmio = pp->ctl_block;
fbbb262d
RH
896 u16 status = readw(mmio + NV_ADMA_STAT);
897 u32 notifier = readl(mmio + NV_ADMA_NOTIFIER);
898 u32 notifier_error = readl(mmio + NV_ADMA_NOTIFIER_ERROR);
0d5ff566 899 void __iomem *dma_stat_addr = ap->ioaddr.bmdma_addr + ATA_DMA_STATUS;
fbbb262d
RH
900
901 /* clear ADMA status */
902 writew(status, mmio + NV_ADMA_STAT);
903 writel(notifier | notifier_error,
cdf56bcf 904 pp->notifier_clear_block);
fbbb262d
RH
905
906 /** clear legacy status */
0d5ff566 907 iowrite8(ioread8(dma_stat_addr), dma_stat_addr);
fbbb262d
RH
908}
909
f5ecac2d 910static void nv_adma_post_internal_cmd(struct ata_queued_cmd *qc)
fbbb262d 911{
f5ecac2d 912 struct nv_adma_port_priv *pp = qc->ap->private_data;
fbbb262d 913
f5ecac2d
RH
914 if(pp->flags & NV_ADMA_PORT_REGISTER_MODE)
915 ata_bmdma_post_internal_cmd(qc);
fbbb262d
RH
916}
917
918static int nv_adma_port_start(struct ata_port *ap)
919{
920 struct device *dev = ap->host->dev;
921 struct nv_adma_port_priv *pp;
922 int rc;
923 void *mem;
924 dma_addr_t mem_dma;
cdf56bcf 925 void __iomem *mmio;
fbbb262d
RH
926 u16 tmp;
927
928 VPRINTK("ENTER\n");
929
930 rc = ata_port_start(ap);
931 if (rc)
932 return rc;
933
24dc5f33
TH
934 pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL);
935 if (!pp)
936 return -ENOMEM;
fbbb262d 937
0d5ff566 938 mmio = ap->host->iomap[NV_MMIO_BAR] + NV_ADMA_PORT +
cdf56bcf
RH
939 ap->port_no * NV_ADMA_PORT_SIZE;
940 pp->ctl_block = mmio;
0d5ff566 941 pp->gen_block = ap->host->iomap[NV_MMIO_BAR] + NV_ADMA_GEN;
cdf56bcf
RH
942 pp->notifier_clear_block = pp->gen_block +
943 NV_ADMA_NOTIFIER_CLEAR + (4 * ap->port_no);
944
24dc5f33
TH
945 mem = dmam_alloc_coherent(dev, NV_ADMA_PORT_PRIV_DMA_SZ,
946 &mem_dma, GFP_KERNEL);
947 if (!mem)
948 return -ENOMEM;
fbbb262d
RH
949 memset(mem, 0, NV_ADMA_PORT_PRIV_DMA_SZ);
950
951 /*
952 * First item in chunk of DMA memory:
953 * 128-byte command parameter block (CPB)
954 * one for each command tag
955 */
956 pp->cpb = mem;
957 pp->cpb_dma = mem_dma;
958
959 writel(mem_dma & 0xFFFFFFFF, mmio + NV_ADMA_CPB_BASE_LOW);
960 writel((mem_dma >> 16 ) >> 16, mmio + NV_ADMA_CPB_BASE_HIGH);
961
962 mem += NV_ADMA_MAX_CPBS * NV_ADMA_CPB_SZ;
963 mem_dma += NV_ADMA_MAX_CPBS * NV_ADMA_CPB_SZ;
964
965 /*
966 * Second item: block of ADMA_SGTBL_LEN s/g entries
967 */
968 pp->aprd = mem;
969 pp->aprd_dma = mem_dma;
970
971 ap->private_data = pp;
972
973 /* clear any outstanding interrupt conditions */
974 writew(0xffff, mmio + NV_ADMA_STAT);
975
976 /* initialize port variables */
977 pp->flags = NV_ADMA_PORT_REGISTER_MODE;
978
979 /* clear CPB fetch count */
980 writew(0, mmio + NV_ADMA_CPB_COUNT);
981
cdf56bcf 982 /* clear GO for register mode, enable interrupt */
fbbb262d 983 tmp = readw(mmio + NV_ADMA_CTL);
5ce0cf6f
RH
984 writew( (tmp & ~NV_ADMA_CTL_GO) | NV_ADMA_CTL_AIEN |
985 NV_ADMA_CTL_HOTPLUG_IEN, mmio + NV_ADMA_CTL);
fbbb262d
RH
986
987 tmp = readw(mmio + NV_ADMA_CTL);
988 writew(tmp | NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 989 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
fbbb262d
RH
990 udelay(1);
991 writew(tmp & ~NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 992 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
fbbb262d
RH
993
994 return 0;
fbbb262d
RH
995}
996
997static void nv_adma_port_stop(struct ata_port *ap)
998{
fbbb262d 999 struct nv_adma_port_priv *pp = ap->private_data;
cdf56bcf 1000 void __iomem *mmio = pp->ctl_block;
fbbb262d
RH
1001
1002 VPRINTK("ENTER\n");
fbbb262d 1003 writew(0, mmio + NV_ADMA_CTL);
fbbb262d
RH
1004}
1005
cdf56bcf
RH
1006static int nv_adma_port_suspend(struct ata_port *ap, pm_message_t mesg)
1007{
1008 struct nv_adma_port_priv *pp = ap->private_data;
1009 void __iomem *mmio = pp->ctl_block;
1010
1011 /* Go to register mode - clears GO */
1012 nv_adma_register_mode(ap);
1013
1014 /* clear CPB fetch count */
1015 writew(0, mmio + NV_ADMA_CPB_COUNT);
1016
1017 /* disable interrupt, shut down port */
1018 writew(0, mmio + NV_ADMA_CTL);
1019
1020 return 0;
1021}
1022
1023static int nv_adma_port_resume(struct ata_port *ap)
1024{
1025 struct nv_adma_port_priv *pp = ap->private_data;
1026 void __iomem *mmio = pp->ctl_block;
1027 u16 tmp;
1028
1029 /* set CPB block location */
1030 writel(pp->cpb_dma & 0xFFFFFFFF, mmio + NV_ADMA_CPB_BASE_LOW);
1031 writel((pp->cpb_dma >> 16 ) >> 16, mmio + NV_ADMA_CPB_BASE_HIGH);
1032
1033 /* clear any outstanding interrupt conditions */
1034 writew(0xffff, mmio + NV_ADMA_STAT);
1035
1036 /* initialize port variables */
1037 pp->flags |= NV_ADMA_PORT_REGISTER_MODE;
1038
1039 /* clear CPB fetch count */
1040 writew(0, mmio + NV_ADMA_CPB_COUNT);
1041
1042 /* clear GO for register mode, enable interrupt */
1043 tmp = readw(mmio + NV_ADMA_CTL);
5ce0cf6f
RH
1044 writew( (tmp & ~NV_ADMA_CTL_GO) | NV_ADMA_CTL_AIEN |
1045 NV_ADMA_CTL_HOTPLUG_IEN, mmio + NV_ADMA_CTL);
cdf56bcf
RH
1046
1047 tmp = readw(mmio + NV_ADMA_CTL);
1048 writew(tmp | NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 1049 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
cdf56bcf
RH
1050 udelay(1);
1051 writew(tmp & ~NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 1052 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
cdf56bcf
RH
1053
1054 return 0;
1055}
fbbb262d
RH
1056
1057static void nv_adma_setup_port(struct ata_probe_ent *probe_ent, unsigned int port)
1058{
0d5ff566 1059 void __iomem *mmio = probe_ent->iomap[NV_MMIO_BAR];
fbbb262d
RH
1060 struct ata_ioports *ioport = &probe_ent->port[port];
1061
1062 VPRINTK("ENTER\n");
1063
1064 mmio += NV_ADMA_PORT + port * NV_ADMA_PORT_SIZE;
1065
0d5ff566
TH
1066 ioport->cmd_addr = mmio;
1067 ioport->data_addr = mmio + (ATA_REG_DATA * 4);
fbbb262d 1068 ioport->error_addr =
0d5ff566
TH
1069 ioport->feature_addr = mmio + (ATA_REG_ERR * 4);
1070 ioport->nsect_addr = mmio + (ATA_REG_NSECT * 4);
1071 ioport->lbal_addr = mmio + (ATA_REG_LBAL * 4);
1072 ioport->lbam_addr = mmio + (ATA_REG_LBAM * 4);
1073 ioport->lbah_addr = mmio + (ATA_REG_LBAH * 4);
1074 ioport->device_addr = mmio + (ATA_REG_DEVICE * 4);
fbbb262d 1075 ioport->status_addr =
0d5ff566 1076 ioport->command_addr = mmio + (ATA_REG_STATUS * 4);
fbbb262d 1077 ioport->altstatus_addr =
0d5ff566 1078 ioport->ctl_addr = mmio + 0x20;
fbbb262d
RH
1079}
1080
1081static int nv_adma_host_init(struct ata_probe_ent *probe_ent)
1082{
1083 struct pci_dev *pdev = to_pci_dev(probe_ent->dev);
1084 unsigned int i;
1085 u32 tmp32;
1086
1087 VPRINTK("ENTER\n");
1088
1089 /* enable ADMA on the ports */
1090 pci_read_config_dword(pdev, NV_MCP_SATA_CFG_20, &tmp32);
1091 tmp32 |= NV_MCP_SATA_CFG_20_PORT0_EN |
1092 NV_MCP_SATA_CFG_20_PORT0_PWB_EN |
1093 NV_MCP_SATA_CFG_20_PORT1_EN |
1094 NV_MCP_SATA_CFG_20_PORT1_PWB_EN;
1095
1096 pci_write_config_dword(pdev, NV_MCP_SATA_CFG_20, tmp32);
1097
1098 for (i = 0; i < probe_ent->n_ports; i++)
1099 nv_adma_setup_port(probe_ent, i);
1100
fbbb262d
RH
1101 return 0;
1102}
1103
1104static void nv_adma_fill_aprd(struct ata_queued_cmd *qc,
1105 struct scatterlist *sg,
1106 int idx,
1107 struct nv_adma_prd *aprd)
1108{
41949ed5 1109 u8 flags = 0;
fbbb262d
RH
1110 if (qc->tf.flags & ATA_TFLAG_WRITE)
1111 flags |= NV_APRD_WRITE;
1112 if (idx == qc->n_elem - 1)
1113 flags |= NV_APRD_END;
1114 else if (idx != 4)
1115 flags |= NV_APRD_CONT;
1116
1117 aprd->addr = cpu_to_le64(((u64)sg_dma_address(sg)));
1118 aprd->len = cpu_to_le32(((u32)sg_dma_len(sg))); /* len in bytes */
2dec7555 1119 aprd->flags = flags;
41949ed5 1120 aprd->packet_len = 0;
fbbb262d
RH
1121}
1122
1123static void nv_adma_fill_sg(struct ata_queued_cmd *qc, struct nv_adma_cpb *cpb)
1124{
1125 struct nv_adma_port_priv *pp = qc->ap->private_data;
1126 unsigned int idx;
1127 struct nv_adma_prd *aprd;
1128 struct scatterlist *sg;
1129
1130 VPRINTK("ENTER\n");
1131
1132 idx = 0;
1133
1134 ata_for_each_sg(sg, qc) {
1135 aprd = (idx < 5) ? &cpb->aprd[idx] : &pp->aprd[NV_ADMA_SGTBL_LEN * qc->tag + (idx-5)];
1136 nv_adma_fill_aprd(qc, sg, idx, aprd);
1137 idx++;
1138 }
1139 if (idx > 5)
1140 cpb->next_aprd = cpu_to_le64(((u64)(pp->aprd_dma + NV_ADMA_SGTBL_SZ * qc->tag)));
41949ed5
RH
1141 else
1142 cpb->next_aprd = cpu_to_le64(0);
fbbb262d
RH
1143}
1144
382a6652
RH
1145static int nv_adma_use_reg_mode(struct ata_queued_cmd *qc)
1146{
1147 struct nv_adma_port_priv *pp = qc->ap->private_data;
1148
1149 /* ADMA engine can only be used for non-ATAPI DMA commands,
1150 or interrupt-driven no-data commands. */
1151 if((pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) ||
1152 (qc->tf.flags & ATA_TFLAG_POLLING))
1153 return 1;
1154
1155 if((qc->flags & ATA_QCFLAG_DMAMAP) ||
1156 (qc->tf.protocol == ATA_PROT_NODATA))
1157 return 0;
1158
1159 return 1;
1160}
1161
fbbb262d
RH
1162static void nv_adma_qc_prep(struct ata_queued_cmd *qc)
1163{
1164 struct nv_adma_port_priv *pp = qc->ap->private_data;
1165 struct nv_adma_cpb *cpb = &pp->cpb[qc->tag];
1166 u8 ctl_flags = NV_CPB_CTL_CPB_VALID |
fbbb262d
RH
1167 NV_CPB_CTL_IEN;
1168
382a6652 1169 if (nv_adma_use_reg_mode(qc)) {
2dec7555 1170 nv_adma_register_mode(qc->ap);
fbbb262d
RH
1171 ata_qc_prep(qc);
1172 return;
1173 }
1174
41949ed5
RH
1175 cpb->resp_flags = NV_CPB_RESP_DONE;
1176 wmb();
1177 cpb->ctl_flags = 0;
1178 wmb();
fbbb262d
RH
1179
1180 cpb->len = 3;
1181 cpb->tag = qc->tag;
1182 cpb->next_cpb_idx = 0;
1183
1184 /* turn on NCQ flags for NCQ commands */
1185 if (qc->tf.protocol == ATA_PROT_NCQ)
1186 ctl_flags |= NV_CPB_CTL_QUEUE | NV_CPB_CTL_FPDMA;
1187
cdf56bcf
RH
1188 VPRINTK("qc->flags = 0x%lx\n", qc->flags);
1189
fbbb262d
RH
1190 nv_adma_tf_to_cpb(&qc->tf, cpb->tf);
1191
382a6652
RH
1192 if(qc->flags & ATA_QCFLAG_DMAMAP) {
1193 nv_adma_fill_sg(qc, cpb);
1194 ctl_flags |= NV_CPB_CTL_APRD_VALID;
1195 } else
1196 memset(&cpb->aprd[0], 0, sizeof(struct nv_adma_prd) * 5);
fbbb262d
RH
1197
1198 /* Be paranoid and don't let the device see NV_CPB_CTL_CPB_VALID until we are
1199 finished filling in all of the contents */
1200 wmb();
1201 cpb->ctl_flags = ctl_flags;
41949ed5
RH
1202 wmb();
1203 cpb->resp_flags = 0;
fbbb262d
RH
1204}
1205
1206static unsigned int nv_adma_qc_issue(struct ata_queued_cmd *qc)
1207{
2dec7555 1208 struct nv_adma_port_priv *pp = qc->ap->private_data;
cdf56bcf 1209 void __iomem *mmio = pp->ctl_block;
5e5c74a5 1210 int curr_ncq = (qc->tf.protocol == ATA_PROT_NCQ);
fbbb262d
RH
1211
1212 VPRINTK("ENTER\n");
1213
382a6652 1214 if (nv_adma_use_reg_mode(qc)) {
fbbb262d 1215 /* use ATA register mode */
382a6652 1216 VPRINTK("using ATA register mode: 0x%lx\n", qc->flags);
fbbb262d
RH
1217 nv_adma_register_mode(qc->ap);
1218 return ata_qc_issue_prot(qc);
1219 } else
1220 nv_adma_mode(qc->ap);
1221
1222 /* write append register, command tag in lower 8 bits
1223 and (number of cpbs to append -1) in top 8 bits */
1224 wmb();
5e5c74a5
RH
1225
1226 if(curr_ncq != pp->last_issue_ncq) {
1227 /* Seems to need some delay before switching between NCQ and non-NCQ
1228 commands, else we get command timeouts and such. */
1229 udelay(20);
1230 pp->last_issue_ncq = curr_ncq;
1231 }
1232
fbbb262d
RH
1233 writew(qc->tag, mmio + NV_ADMA_APPEND);
1234
1235 DPRINTK("Issued tag %u\n",qc->tag);
1236
1237 return 0;
1238}
1239
7d12e780 1240static irqreturn_t nv_generic_interrupt(int irq, void *dev_instance)
1da177e4 1241{
cca3974e 1242 struct ata_host *host = dev_instance;
1da177e4
LT
1243 unsigned int i;
1244 unsigned int handled = 0;
1245 unsigned long flags;
1246
cca3974e 1247 spin_lock_irqsave(&host->lock, flags);
1da177e4 1248
cca3974e 1249 for (i = 0; i < host->n_ports; i++) {
1da177e4
LT
1250 struct ata_port *ap;
1251
cca3974e 1252 ap = host->ports[i];
c1389503 1253 if (ap &&
029f5468 1254 !(ap->flags & ATA_FLAG_DISABLED)) {
1da177e4
LT
1255 struct ata_queued_cmd *qc;
1256
1257 qc = ata_qc_from_tag(ap, ap->active_tag);
e50362ec 1258 if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)))
1da177e4 1259 handled += ata_host_intr(ap, qc);
b887030a
AC
1260 else
1261 // No request pending? Clear interrupt status
1262 // anyway, in case there's one pending.
1263 ap->ops->check_status(ap);
1da177e4
LT
1264 }
1265
1266 }
1267
cca3974e 1268 spin_unlock_irqrestore(&host->lock, flags);
1da177e4
LT
1269
1270 return IRQ_RETVAL(handled);
1271}
1272
cca3974e 1273static irqreturn_t nv_do_interrupt(struct ata_host *host, u8 irq_stat)
ada364e8
TH
1274{
1275 int i, handled = 0;
1276
cca3974e
JG
1277 for (i = 0; i < host->n_ports; i++) {
1278 struct ata_port *ap = host->ports[i];
ada364e8
TH
1279
1280 if (ap && !(ap->flags & ATA_FLAG_DISABLED))
1281 handled += nv_host_intr(ap, irq_stat);
1282
1283 irq_stat >>= NV_INT_PORT_SHIFT;
1284 }
1285
1286 return IRQ_RETVAL(handled);
1287}
1288
7d12e780 1289static irqreturn_t nv_nf2_interrupt(int irq, void *dev_instance)
ada364e8 1290{
cca3974e 1291 struct ata_host *host = dev_instance;
ada364e8
TH
1292 u8 irq_stat;
1293 irqreturn_t ret;
1294
cca3974e 1295 spin_lock(&host->lock);
0d5ff566 1296 irq_stat = ioread8(host->ports[0]->ioaddr.scr_addr + NV_INT_STATUS);
cca3974e
JG
1297 ret = nv_do_interrupt(host, irq_stat);
1298 spin_unlock(&host->lock);
ada364e8
TH
1299
1300 return ret;
1301}
1302
7d12e780 1303static irqreturn_t nv_ck804_interrupt(int irq, void *dev_instance)
ada364e8 1304{
cca3974e 1305 struct ata_host *host = dev_instance;
ada364e8
TH
1306 u8 irq_stat;
1307 irqreturn_t ret;
1308
cca3974e 1309 spin_lock(&host->lock);
0d5ff566 1310 irq_stat = readb(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804);
cca3974e
JG
1311 ret = nv_do_interrupt(host, irq_stat);
1312 spin_unlock(&host->lock);
ada364e8
TH
1313
1314 return ret;
1315}
1316
1da177e4
LT
1317static u32 nv_scr_read (struct ata_port *ap, unsigned int sc_reg)
1318{
1da177e4
LT
1319 if (sc_reg > SCR_CONTROL)
1320 return 0xffffffffU;
1321
0d5ff566 1322 return ioread32(ap->ioaddr.scr_addr + (sc_reg * 4));
1da177e4
LT
1323}
1324
1325static void nv_scr_write (struct ata_port *ap, unsigned int sc_reg, u32 val)
1326{
1da177e4
LT
1327 if (sc_reg > SCR_CONTROL)
1328 return;
1329
0d5ff566 1330 iowrite32(val, ap->ioaddr.scr_addr + (sc_reg * 4));
1da177e4
LT
1331}
1332
39f87582
TH
1333static void nv_nf2_freeze(struct ata_port *ap)
1334{
0d5ff566 1335 void __iomem *scr_addr = ap->host->ports[0]->ioaddr.scr_addr;
39f87582
TH
1336 int shift = ap->port_no * NV_INT_PORT_SHIFT;
1337 u8 mask;
1338
0d5ff566 1339 mask = ioread8(scr_addr + NV_INT_ENABLE);
39f87582 1340 mask &= ~(NV_INT_ALL << shift);
0d5ff566 1341 iowrite8(mask, scr_addr + NV_INT_ENABLE);
39f87582
TH
1342}
1343
1344static void nv_nf2_thaw(struct ata_port *ap)
1345{
0d5ff566 1346 void __iomem *scr_addr = ap->host->ports[0]->ioaddr.scr_addr;
39f87582
TH
1347 int shift = ap->port_no * NV_INT_PORT_SHIFT;
1348 u8 mask;
1349
0d5ff566 1350 iowrite8(NV_INT_ALL << shift, scr_addr + NV_INT_STATUS);
39f87582 1351
0d5ff566 1352 mask = ioread8(scr_addr + NV_INT_ENABLE);
39f87582 1353 mask |= (NV_INT_MASK << shift);
0d5ff566 1354 iowrite8(mask, scr_addr + NV_INT_ENABLE);
39f87582
TH
1355}
1356
1357static void nv_ck804_freeze(struct ata_port *ap)
1358{
0d5ff566 1359 void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
39f87582
TH
1360 int shift = ap->port_no * NV_INT_PORT_SHIFT;
1361 u8 mask;
1362
1363 mask = readb(mmio_base + NV_INT_ENABLE_CK804);
1364 mask &= ~(NV_INT_ALL << shift);
1365 writeb(mask, mmio_base + NV_INT_ENABLE_CK804);
1366}
1367
1368static void nv_ck804_thaw(struct ata_port *ap)
1369{
0d5ff566 1370 void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
39f87582
TH
1371 int shift = ap->port_no * NV_INT_PORT_SHIFT;
1372 u8 mask;
1373
1374 writeb(NV_INT_ALL << shift, mmio_base + NV_INT_STATUS_CK804);
1375
1376 mask = readb(mmio_base + NV_INT_ENABLE_CK804);
1377 mask |= (NV_INT_MASK << shift);
1378 writeb(mask, mmio_base + NV_INT_ENABLE_CK804);
1379}
1380
1381static int nv_hardreset(struct ata_port *ap, unsigned int *class)
1382{
1383 unsigned int dummy;
1384
1385 /* SATA hardreset fails to retrieve proper device signature on
1386 * some controllers. Don't classify on hardreset. For more
1387 * info, see http://bugme.osdl.org/show_bug.cgi?id=3352
1388 */
1389 return sata_std_hardreset(ap, &dummy);
1390}
1391
1392static void nv_error_handler(struct ata_port *ap)
1393{
1394 ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset,
1395 nv_hardreset, ata_std_postreset);
1396}
1397
fbbb262d
RH
1398static void nv_adma_error_handler(struct ata_port *ap)
1399{
1400 struct nv_adma_port_priv *pp = ap->private_data;
1401 if(!(pp->flags & NV_ADMA_PORT_REGISTER_MODE)) {
cdf56bcf 1402 void __iomem *mmio = pp->ctl_block;
fbbb262d
RH
1403 int i;
1404 u16 tmp;
2cb27853
RH
1405
1406 if(ata_tag_valid(ap->active_tag) || ap->sactive) {
1407 u32 notifier = readl(mmio + NV_ADMA_NOTIFIER);
1408 u32 notifier_error = readl(mmio + NV_ADMA_NOTIFIER_ERROR);
1409 u32 gen_ctl = readl(pp->gen_block + NV_ADMA_GEN_CTL);
1410 u32 status = readw(mmio + NV_ADMA_STAT);
08af7414
RH
1411 u8 cpb_count = readb(mmio + NV_ADMA_CPB_COUNT);
1412 u8 next_cpb_idx = readb(mmio + NV_ADMA_NEXT_CPB_IDX);
2cb27853
RH
1413
1414 ata_port_printk(ap, KERN_ERR, "EH in ADMA mode, notifier 0x%X "
08af7414
RH
1415 "notifier_error 0x%X gen_ctl 0x%X status 0x%X "
1416 "next cpb count 0x%X next cpb idx 0x%x\n",
1417 notifier, notifier_error, gen_ctl, status,
1418 cpb_count, next_cpb_idx);
2cb27853
RH
1419
1420 for( i=0;i<NV_ADMA_MAX_CPBS;i++) {
1421 struct nv_adma_cpb *cpb = &pp->cpb[i];
1422 if( (ata_tag_valid(ap->active_tag) && i == ap->active_tag) ||
1423 ap->sactive & (1 << i) )
1424 ata_port_printk(ap, KERN_ERR,
1425 "CPB %d: ctl_flags 0x%x, resp_flags 0x%x\n",
1426 i, cpb->ctl_flags, cpb->resp_flags);
1427 }
1428 }
fbbb262d 1429
fbbb262d
RH
1430 /* Push us back into port register mode for error handling. */
1431 nv_adma_register_mode(ap);
1432
fbbb262d
RH
1433 /* Mark all of the CPBs as invalid to prevent them from being executed */
1434 for( i=0;i<NV_ADMA_MAX_CPBS;i++)
1435 pp->cpb[i].ctl_flags &= ~NV_CPB_CTL_CPB_VALID;
1436
1437 /* clear CPB fetch count */
1438 writew(0, mmio + NV_ADMA_CPB_COUNT);
1439
1440 /* Reset channel */
1441 tmp = readw(mmio + NV_ADMA_CTL);
1442 writew(tmp | NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 1443 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
fbbb262d
RH
1444 udelay(1);
1445 writew(tmp & ~NV_ADMA_CTL_CHANNEL_RESET, mmio + NV_ADMA_CTL);
5ce0cf6f 1446 readw( mmio + NV_ADMA_CTL ); /* flush posted write */
fbbb262d
RH
1447 }
1448
1449 ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset,
1450 nv_hardreset, ata_std_postreset);
1451}
1452
1da177e4
LT
1453static int nv_init_one (struct pci_dev *pdev, const struct pci_device_id *ent)
1454{
1455 static int printed_version = 0;
29da9f6d 1456 struct ata_port_info *ppi[2];
1da177e4 1457 struct ata_probe_ent *probe_ent;
cdf56bcf 1458 struct nv_host_priv *hpriv;
1da177e4
LT
1459 int rc;
1460 u32 bar;
0d5ff566 1461 void __iomem *base;
fbbb262d
RH
1462 unsigned long type = ent->driver_data;
1463 int mask_set = 0;
1da177e4
LT
1464
1465 // Make sure this is a SATA controller by counting the number of bars
1466 // (NVIDIA SATA controllers will always have six bars). Otherwise,
1467 // it's an IDE controller and we ignore it.
1468 for (bar=0; bar<6; bar++)
1469 if (pci_resource_start(pdev, bar) == 0)
1470 return -ENODEV;
1471
cdf56bcf 1472 if (!printed_version++)
a9524a76 1473 dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n");
1da177e4 1474
24dc5f33 1475 rc = pcim_enable_device(pdev);
1da177e4 1476 if (rc)
24dc5f33 1477 return rc;
1da177e4
LT
1478
1479 rc = pci_request_regions(pdev, DRV_NAME);
1480 if (rc) {
24dc5f33
TH
1481 pcim_pin_device(pdev);
1482 return rc;
1da177e4
LT
1483 }
1484
fbbb262d
RH
1485 if(type >= CK804 && adma_enabled) {
1486 dev_printk(KERN_NOTICE, &pdev->dev, "Using ADMA mode\n");
1487 type = ADMA;
1488 if(!pci_set_dma_mask(pdev, DMA_64BIT_MASK) &&
1489 !pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK))
1490 mask_set = 1;
1491 }
1492
1493 if(!mask_set) {
1494 rc = pci_set_dma_mask(pdev, ATA_DMA_MASK);
1495 if (rc)
24dc5f33 1496 return rc;
fbbb262d
RH
1497 rc = pci_set_consistent_dma_mask(pdev, ATA_DMA_MASK);
1498 if (rc)
24dc5f33 1499 return rc;
fbbb262d 1500 }
1da177e4
LT
1501
1502 rc = -ENOMEM;
1503
24dc5f33 1504 hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL);
cdf56bcf 1505 if (!hpriv)
24dc5f33 1506 return -ENOMEM;
cdf56bcf 1507
fbbb262d 1508 ppi[0] = ppi[1] = &nv_port_info[type];
29da9f6d 1509 probe_ent = ata_pci_init_native_mode(pdev, ppi, ATA_PORT_PRIMARY | ATA_PORT_SECONDARY);
1da177e4 1510 if (!probe_ent)
24dc5f33
TH
1511 return -ENOMEM;
1512
0d5ff566 1513 if (!pcim_iomap(pdev, NV_MMIO_BAR, 0))
24dc5f33 1514 return -EIO;
0d5ff566 1515 probe_ent->iomap = pcim_iomap_table(pdev);
1da177e4 1516
cdf56bcf
RH
1517 probe_ent->private_data = hpriv;
1518 hpriv->type = type;
1da177e4 1519
0d5ff566 1520 base = probe_ent->iomap[NV_MMIO_BAR];
02cbd926
JG
1521 probe_ent->port[0].scr_addr = base + NV_PORT0_SCR_REG_OFFSET;
1522 probe_ent->port[1].scr_addr = base + NV_PORT1_SCR_REG_OFFSET;
1da177e4 1523
ada364e8 1524 /* enable SATA space for CK804 */
fbbb262d 1525 if (type >= CK804) {
ada364e8
TH
1526 u8 regval;
1527
1528 pci_read_config_byte(pdev, NV_MCP_SATA_CFG_20, &regval);
1529 regval |= NV_MCP_SATA_CFG_20_SATA_SPACE_EN;
1530 pci_write_config_byte(pdev, NV_MCP_SATA_CFG_20, regval);
1531 }
1532
1da177e4
LT
1533 pci_set_master(pdev);
1534
fbbb262d
RH
1535 if (type == ADMA) {
1536 rc = nv_adma_host_init(probe_ent);
1537 if (rc)
24dc5f33 1538 return rc;
fbbb262d
RH
1539 }
1540
1da177e4
LT
1541 rc = ata_device_add(probe_ent);
1542 if (rc != NV_PORTS)
24dc5f33 1543 return -ENODEV;
1da177e4 1544
24dc5f33 1545 devm_kfree(&pdev->dev, probe_ent);
1da177e4 1546 return 0;
1da177e4
LT
1547}
1548
cdf56bcf
RH
1549static void nv_remove_one (struct pci_dev *pdev)
1550{
1551 struct ata_host *host = dev_get_drvdata(&pdev->dev);
1552 struct nv_host_priv *hpriv = host->private_data;
1553
1554 ata_pci_remove_one(pdev);
1555 kfree(hpriv);
1556}
1557
1558static int nv_pci_device_resume(struct pci_dev *pdev)
1559{
1560 struct ata_host *host = dev_get_drvdata(&pdev->dev);
1561 struct nv_host_priv *hpriv = host->private_data;
ce053fa8 1562 int rc;
cdf56bcf 1563
ce053fa8
RH
1564 rc = ata_pci_device_do_resume(pdev);
1565 if(rc)
1566 return rc;
cdf56bcf
RH
1567
1568 if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
1569 if(hpriv->type >= CK804) {
1570 u8 regval;
1571
1572 pci_read_config_byte(pdev, NV_MCP_SATA_CFG_20, &regval);
1573 regval |= NV_MCP_SATA_CFG_20_SATA_SPACE_EN;
1574 pci_write_config_byte(pdev, NV_MCP_SATA_CFG_20, regval);
1575 }
1576 if(hpriv->type == ADMA) {
1577 u32 tmp32;
1578 struct nv_adma_port_priv *pp;
1579 /* enable/disable ADMA on the ports appropriately */
1580 pci_read_config_dword(pdev, NV_MCP_SATA_CFG_20, &tmp32);
1581
1582 pp = host->ports[0]->private_data;
1583 if(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE)
1584 tmp32 &= ~(NV_MCP_SATA_CFG_20_PORT0_EN |
1585 NV_MCP_SATA_CFG_20_PORT0_PWB_EN);
1586 else
1587 tmp32 |= (NV_MCP_SATA_CFG_20_PORT0_EN |
1588 NV_MCP_SATA_CFG_20_PORT0_PWB_EN);
1589 pp = host->ports[1]->private_data;
1590 if(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE)
1591 tmp32 &= ~(NV_MCP_SATA_CFG_20_PORT1_EN |
1592 NV_MCP_SATA_CFG_20_PORT1_PWB_EN);
1593 else
1594 tmp32 |= (NV_MCP_SATA_CFG_20_PORT1_EN |
1595 NV_MCP_SATA_CFG_20_PORT1_PWB_EN);
1596
1597 pci_write_config_dword(pdev, NV_MCP_SATA_CFG_20, tmp32);
1598 }
1599 }
1600
1601 ata_host_resume(host);
1602
1603 return 0;
1604}
1605
cca3974e 1606static void nv_ck804_host_stop(struct ata_host *host)
ada364e8 1607{
cca3974e 1608 struct pci_dev *pdev = to_pci_dev(host->dev);
ada364e8
TH
1609 u8 regval;
1610
1611 /* disable SATA space for CK804 */
1612 pci_read_config_byte(pdev, NV_MCP_SATA_CFG_20, &regval);
1613 regval &= ~NV_MCP_SATA_CFG_20_SATA_SPACE_EN;
1614 pci_write_config_byte(pdev, NV_MCP_SATA_CFG_20, regval);
ada364e8
TH
1615}
1616
fbbb262d
RH
1617static void nv_adma_host_stop(struct ata_host *host)
1618{
1619 struct pci_dev *pdev = to_pci_dev(host->dev);
fbbb262d
RH
1620 u32 tmp32;
1621
fbbb262d
RH
1622 /* disable ADMA on the ports */
1623 pci_read_config_dword(pdev, NV_MCP_SATA_CFG_20, &tmp32);
1624 tmp32 &= ~(NV_MCP_SATA_CFG_20_PORT0_EN |
1625 NV_MCP_SATA_CFG_20_PORT0_PWB_EN |
1626 NV_MCP_SATA_CFG_20_PORT1_EN |
1627 NV_MCP_SATA_CFG_20_PORT1_PWB_EN);
1628
1629 pci_write_config_dword(pdev, NV_MCP_SATA_CFG_20, tmp32);
1630
1631 nv_ck804_host_stop(host);
1632}
1633
1da177e4
LT
1634static int __init nv_init(void)
1635{
b7887196 1636 return pci_register_driver(&nv_pci_driver);
1da177e4
LT
1637}
1638
1639static void __exit nv_exit(void)
1640{
1641 pci_unregister_driver(&nv_pci_driver);
1642}
1643
1644module_init(nv_init);
1645module_exit(nv_exit);
fbbb262d
RH
1646module_param_named(adma, adma_enabled, bool, 0444);
1647MODULE_PARM_DESC(adma, "Enable use of ADMA (Default: true)");