]>
Commit | Line | Data |
---|---|---|
1da177e4 | 1 | /* |
1da177e4 LT |
2 | * Copyright (C) 2000-2002 Andre Hedrick <andre@linux-ide.org> |
3 | * Copyright (C) 2003 Red Hat <alan@redhat.com> | |
4 | * | |
5 | */ | |
6 | ||
1da177e4 LT |
7 | #include <linux/module.h> |
8 | #include <linux/types.h> | |
9 | #include <linux/string.h> | |
10 | #include <linux/kernel.h> | |
11 | #include <linux/timer.h> | |
12 | #include <linux/mm.h> | |
13 | #include <linux/interrupt.h> | |
14 | #include <linux/major.h> | |
15 | #include <linux/errno.h> | |
16 | #include <linux/genhd.h> | |
17 | #include <linux/blkpg.h> | |
18 | #include <linux/slab.h> | |
19 | #include <linux/pci.h> | |
20 | #include <linux/delay.h> | |
21 | #include <linux/hdreg.h> | |
22 | #include <linux/ide.h> | |
23 | #include <linux/bitops.h> | |
1e86240f | 24 | #include <linux/nmi.h> |
1da177e4 LT |
25 | |
26 | #include <asm/byteorder.h> | |
27 | #include <asm/irq.h> | |
28 | #include <asm/uaccess.h> | |
29 | #include <asm/io.h> | |
30 | ||
31 | /* | |
32 | * Conventional PIO operations for ATA devices | |
33 | */ | |
34 | ||
35 | static u8 ide_inb (unsigned long port) | |
36 | { | |
37 | return (u8) inb(port); | |
38 | } | |
39 | ||
1da177e4 LT |
40 | static void ide_outb (u8 val, unsigned long port) |
41 | { | |
42 | outb(val, port); | |
43 | } | |
44 | ||
1da177e4 LT |
45 | /* |
46 | * MMIO operations, typically used for SATA controllers | |
47 | */ | |
48 | ||
49 | static u8 ide_mm_inb (unsigned long port) | |
50 | { | |
51 | return (u8) readb((void __iomem *) port); | |
52 | } | |
53 | ||
1da177e4 LT |
54 | static void ide_mm_outb (u8 value, unsigned long port) |
55 | { | |
56 | writeb(value, (void __iomem *) port); | |
57 | } | |
58 | ||
1da177e4 LT |
59 | void SELECT_DRIVE (ide_drive_t *drive) |
60 | { | |
23579a2a | 61 | ide_hwif_t *hwif = drive->hwif; |
ac95beed | 62 | const struct ide_port_ops *port_ops = hwif->port_ops; |
40f095f0 | 63 | ide_task_t task; |
23579a2a | 64 | |
ac95beed BZ |
65 | if (port_ops && port_ops->selectproc) |
66 | port_ops->selectproc(drive); | |
23579a2a | 67 | |
40f095f0 BZ |
68 | memset(&task, 0, sizeof(task)); |
69 | task.tf_flags = IDE_TFLAG_OUT_DEVICE; | |
70 | ||
71 | drive->hwif->tf_load(drive, &task); | |
1da177e4 LT |
72 | } |
73 | ||
ed4af48f | 74 | void SELECT_MASK(ide_drive_t *drive, int mask) |
1da177e4 | 75 | { |
ac95beed BZ |
76 | const struct ide_port_ops *port_ops = drive->hwif->port_ops; |
77 | ||
78 | if (port_ops && port_ops->maskproc) | |
79 | port_ops->maskproc(drive, mask); | |
1da177e4 LT |
80 | } |
81 | ||
c6dfa867 BZ |
82 | static void ide_exec_command(ide_hwif_t *hwif, u8 cmd) |
83 | { | |
84 | if (hwif->host_flags & IDE_HFLAG_MMIO) | |
85 | writeb(cmd, (void __iomem *)hwif->io_ports.command_addr); | |
86 | else | |
87 | outb(cmd, hwif->io_ports.command_addr); | |
88 | } | |
89 | ||
b73c7ee2 BZ |
90 | static u8 ide_read_status(ide_hwif_t *hwif) |
91 | { | |
92 | if (hwif->host_flags & IDE_HFLAG_MMIO) | |
93 | return readb((void __iomem *)hwif->io_ports.status_addr); | |
94 | else | |
95 | return inb(hwif->io_ports.status_addr); | |
96 | } | |
97 | ||
1f6d8a0f BZ |
98 | static u8 ide_read_altstatus(ide_hwif_t *hwif) |
99 | { | |
100 | if (hwif->host_flags & IDE_HFLAG_MMIO) | |
101 | return readb((void __iomem *)hwif->io_ports.ctl_addr); | |
102 | else | |
103 | return inb(hwif->io_ports.ctl_addr); | |
104 | } | |
105 | ||
b2f951aa BZ |
106 | static u8 ide_read_sff_dma_status(ide_hwif_t *hwif) |
107 | { | |
108 | if (hwif->host_flags & IDE_HFLAG_MMIO) | |
cab7f8ed | 109 | return readb((void __iomem *)(hwif->dma_base + ATA_DMA_STATUS)); |
b2f951aa | 110 | else |
cab7f8ed | 111 | return inb(hwif->dma_base + ATA_DMA_STATUS); |
b2f951aa BZ |
112 | } |
113 | ||
6e6afb3b BZ |
114 | static void ide_set_irq(ide_hwif_t *hwif, int on) |
115 | { | |
116 | u8 ctl = ATA_DEVCTL_OBS; | |
117 | ||
118 | if (on == 4) { /* hack for SRST */ | |
119 | ctl |= 4; | |
120 | on &= ~4; | |
121 | } | |
122 | ||
123 | ctl |= on ? 0 : 2; | |
124 | ||
125 | if (hwif->host_flags & IDE_HFLAG_MMIO) | |
126 | writeb(ctl, (void __iomem *)hwif->io_ports.ctl_addr); | |
127 | else | |
128 | outb(ctl, hwif->io_ports.ctl_addr); | |
129 | } | |
130 | ||
94cd5b62 | 131 | static void ide_tf_load(ide_drive_t *drive, ide_task_t *task) |
d309e0bb BZ |
132 | { |
133 | ide_hwif_t *hwif = drive->hwif; | |
134 | struct ide_io_ports *io_ports = &hwif->io_ports; | |
135 | struct ide_taskfile *tf = &task->tf; | |
ca545c1e BZ |
136 | void (*tf_outb)(u8 addr, unsigned long port); |
137 | u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; | |
d309e0bb BZ |
138 | u8 HIHI = (task->tf_flags & IDE_TFLAG_LBA48) ? 0xE0 : 0xEF; |
139 | ||
ca545c1e BZ |
140 | if (mmio) |
141 | tf_outb = ide_mm_outb; | |
142 | else | |
143 | tf_outb = ide_outb; | |
144 | ||
d309e0bb BZ |
145 | if (task->tf_flags & IDE_TFLAG_FLAGGED) |
146 | HIHI = 0xFF; | |
147 | ||
ca545c1e BZ |
148 | if (task->tf_flags & IDE_TFLAG_OUT_DATA) { |
149 | u16 data = (tf->hob_data << 8) | tf->data; | |
150 | ||
151 | if (mmio) | |
152 | writew(data, (void __iomem *)io_ports->data_addr); | |
153 | else | |
154 | outw(data, io_ports->data_addr); | |
155 | } | |
d309e0bb BZ |
156 | |
157 | if (task->tf_flags & IDE_TFLAG_OUT_HOB_FEATURE) | |
ca545c1e | 158 | tf_outb(tf->hob_feature, io_ports->feature_addr); |
d309e0bb | 159 | if (task->tf_flags & IDE_TFLAG_OUT_HOB_NSECT) |
ca545c1e | 160 | tf_outb(tf->hob_nsect, io_ports->nsect_addr); |
d309e0bb | 161 | if (task->tf_flags & IDE_TFLAG_OUT_HOB_LBAL) |
ca545c1e | 162 | tf_outb(tf->hob_lbal, io_ports->lbal_addr); |
d309e0bb | 163 | if (task->tf_flags & IDE_TFLAG_OUT_HOB_LBAM) |
ca545c1e | 164 | tf_outb(tf->hob_lbam, io_ports->lbam_addr); |
d309e0bb | 165 | if (task->tf_flags & IDE_TFLAG_OUT_HOB_LBAH) |
ca545c1e | 166 | tf_outb(tf->hob_lbah, io_ports->lbah_addr); |
d309e0bb BZ |
167 | |
168 | if (task->tf_flags & IDE_TFLAG_OUT_FEATURE) | |
ca545c1e | 169 | tf_outb(tf->feature, io_ports->feature_addr); |
d309e0bb | 170 | if (task->tf_flags & IDE_TFLAG_OUT_NSECT) |
ca545c1e | 171 | tf_outb(tf->nsect, io_ports->nsect_addr); |
d309e0bb | 172 | if (task->tf_flags & IDE_TFLAG_OUT_LBAL) |
ca545c1e | 173 | tf_outb(tf->lbal, io_ports->lbal_addr); |
d309e0bb | 174 | if (task->tf_flags & IDE_TFLAG_OUT_LBAM) |
ca545c1e | 175 | tf_outb(tf->lbam, io_ports->lbam_addr); |
d309e0bb | 176 | if (task->tf_flags & IDE_TFLAG_OUT_LBAH) |
ca545c1e | 177 | tf_outb(tf->lbah, io_ports->lbah_addr); |
d309e0bb BZ |
178 | |
179 | if (task->tf_flags & IDE_TFLAG_OUT_DEVICE) | |
ca545c1e BZ |
180 | tf_outb((tf->device & HIHI) | drive->select.all, |
181 | io_ports->device_addr); | |
d309e0bb BZ |
182 | } |
183 | ||
94cd5b62 | 184 | static void ide_tf_read(ide_drive_t *drive, ide_task_t *task) |
d309e0bb BZ |
185 | { |
186 | ide_hwif_t *hwif = drive->hwif; | |
187 | struct ide_io_ports *io_ports = &hwif->io_ports; | |
188 | struct ide_taskfile *tf = &task->tf; | |
ca545c1e BZ |
189 | void (*tf_outb)(u8 addr, unsigned long port); |
190 | u8 (*tf_inb)(unsigned long port); | |
191 | u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; | |
192 | ||
193 | if (mmio) { | |
194 | tf_outb = ide_mm_outb; | |
195 | tf_inb = ide_mm_inb; | |
196 | } else { | |
197 | tf_outb = ide_outb; | |
198 | tf_inb = ide_inb; | |
199 | } | |
d309e0bb BZ |
200 | |
201 | if (task->tf_flags & IDE_TFLAG_IN_DATA) { | |
ca545c1e BZ |
202 | u16 data; |
203 | ||
204 | if (mmio) | |
205 | data = readw((void __iomem *)io_ports->data_addr); | |
206 | else | |
207 | data = inw(io_ports->data_addr); | |
d309e0bb BZ |
208 | |
209 | tf->data = data & 0xff; | |
210 | tf->hob_data = (data >> 8) & 0xff; | |
211 | } | |
212 | ||
213 | /* be sure we're looking at the low order bits */ | |
ff074883 | 214 | tf_outb(ATA_DEVCTL_OBS & ~0x80, io_ports->ctl_addr); |
d309e0bb | 215 | |
92eb4380 BZ |
216 | if (task->tf_flags & IDE_TFLAG_IN_FEATURE) |
217 | tf->feature = tf_inb(io_ports->feature_addr); | |
d309e0bb | 218 | if (task->tf_flags & IDE_TFLAG_IN_NSECT) |
ca545c1e | 219 | tf->nsect = tf_inb(io_ports->nsect_addr); |
d309e0bb | 220 | if (task->tf_flags & IDE_TFLAG_IN_LBAL) |
ca545c1e | 221 | tf->lbal = tf_inb(io_ports->lbal_addr); |
d309e0bb | 222 | if (task->tf_flags & IDE_TFLAG_IN_LBAM) |
ca545c1e | 223 | tf->lbam = tf_inb(io_ports->lbam_addr); |
d309e0bb | 224 | if (task->tf_flags & IDE_TFLAG_IN_LBAH) |
ca545c1e | 225 | tf->lbah = tf_inb(io_ports->lbah_addr); |
d309e0bb | 226 | if (task->tf_flags & IDE_TFLAG_IN_DEVICE) |
ca545c1e | 227 | tf->device = tf_inb(io_ports->device_addr); |
d309e0bb BZ |
228 | |
229 | if (task->tf_flags & IDE_TFLAG_LBA48) { | |
ff074883 | 230 | tf_outb(ATA_DEVCTL_OBS | 0x80, io_ports->ctl_addr); |
d309e0bb BZ |
231 | |
232 | if (task->tf_flags & IDE_TFLAG_IN_HOB_FEATURE) | |
ca545c1e | 233 | tf->hob_feature = tf_inb(io_ports->feature_addr); |
d309e0bb | 234 | if (task->tf_flags & IDE_TFLAG_IN_HOB_NSECT) |
ca545c1e | 235 | tf->hob_nsect = tf_inb(io_ports->nsect_addr); |
d309e0bb | 236 | if (task->tf_flags & IDE_TFLAG_IN_HOB_LBAL) |
ca545c1e | 237 | tf->hob_lbal = tf_inb(io_ports->lbal_addr); |
d309e0bb | 238 | if (task->tf_flags & IDE_TFLAG_IN_HOB_LBAM) |
ca545c1e | 239 | tf->hob_lbam = tf_inb(io_ports->lbam_addr); |
d309e0bb | 240 | if (task->tf_flags & IDE_TFLAG_IN_HOB_LBAH) |
ca545c1e | 241 | tf->hob_lbah = tf_inb(io_ports->lbah_addr); |
d309e0bb BZ |
242 | } |
243 | } | |
244 | ||
1da177e4 LT |
245 | /* |
246 | * Some localbus EIDE interfaces require a special access sequence | |
247 | * when using 32-bit I/O instructions to transfer data. We call this | |
248 | * the "vlb_sync" sequence, which consists of three successive reads | |
249 | * of the sector count register location, with interrupts disabled | |
250 | * to ensure that the reads all happen together. | |
251 | */ | |
22cdd6ce | 252 | static void ata_vlb_sync(unsigned long port) |
1da177e4 | 253 | { |
22cdd6ce BZ |
254 | (void)inb(port); |
255 | (void)inb(port); | |
256 | (void)inb(port); | |
1da177e4 LT |
257 | } |
258 | ||
259 | /* | |
260 | * This is used for most PIO data transfers *from* the IDE interface | |
9567b349 BZ |
261 | * |
262 | * These routines will round up any request for an odd number of bytes, | |
263 | * so if an odd len is specified, be sure that there's at least one | |
264 | * extra byte allocated for the buffer. | |
1da177e4 | 265 | */ |
92d3ab27 | 266 | static void ata_input_data(ide_drive_t *drive, struct request *rq, |
9567b349 | 267 | void *buf, unsigned int len) |
1da177e4 | 268 | { |
4c3032d8 BZ |
269 | ide_hwif_t *hwif = drive->hwif; |
270 | struct ide_io_ports *io_ports = &hwif->io_ports; | |
9567b349 | 271 | unsigned long data_addr = io_ports->data_addr; |
4c3032d8 | 272 | u8 io_32bit = drive->io_32bit; |
16bb69c1 | 273 | u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; |
1da177e4 | 274 | |
9567b349 BZ |
275 | len++; |
276 | ||
1da177e4 | 277 | if (io_32bit) { |
16bb69c1 | 278 | unsigned long uninitialized_var(flags); |
23579a2a | 279 | |
22cdd6ce | 280 | if ((io_32bit & 2) && !mmio) { |
1da177e4 | 281 | local_irq_save(flags); |
22cdd6ce | 282 | ata_vlb_sync(io_ports->nsect_addr); |
16bb69c1 BZ |
283 | } |
284 | ||
285 | if (mmio) | |
286 | __ide_mm_insl((void __iomem *)data_addr, buf, len / 4); | |
287 | else | |
288 | insl(data_addr, buf, len / 4); | |
289 | ||
22cdd6ce | 290 | if ((io_32bit & 2) && !mmio) |
1da177e4 | 291 | local_irq_restore(flags); |
9567b349 | 292 | |
16bb69c1 BZ |
293 | if ((len & 3) >= 2) { |
294 | if (mmio) | |
295 | __ide_mm_insw((void __iomem *)data_addr, | |
296 | (u8 *)buf + (len & ~3), 1); | |
297 | else | |
298 | insw(data_addr, (u8 *)buf + (len & ~3), 1); | |
299 | } | |
300 | } else { | |
301 | if (mmio) | |
302 | __ide_mm_insw((void __iomem *)data_addr, buf, len / 2); | |
303 | else | |
304 | insw(data_addr, buf, len / 2); | |
305 | } | |
1da177e4 LT |
306 | } |
307 | ||
308 | /* | |
309 | * This is used for most PIO data transfers *to* the IDE interface | |
310 | */ | |
92d3ab27 | 311 | static void ata_output_data(ide_drive_t *drive, struct request *rq, |
9567b349 | 312 | void *buf, unsigned int len) |
1da177e4 | 313 | { |
4c3032d8 BZ |
314 | ide_hwif_t *hwif = drive->hwif; |
315 | struct ide_io_ports *io_ports = &hwif->io_ports; | |
9567b349 | 316 | unsigned long data_addr = io_ports->data_addr; |
4c3032d8 | 317 | u8 io_32bit = drive->io_32bit; |
16bb69c1 | 318 | u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; |
1da177e4 LT |
319 | |
320 | if (io_32bit) { | |
16bb69c1 | 321 | unsigned long uninitialized_var(flags); |
23579a2a | 322 | |
22cdd6ce | 323 | if ((io_32bit & 2) && !mmio) { |
1da177e4 | 324 | local_irq_save(flags); |
22cdd6ce | 325 | ata_vlb_sync(io_ports->nsect_addr); |
16bb69c1 BZ |
326 | } |
327 | ||
328 | if (mmio) | |
329 | __ide_mm_outsl((void __iomem *)data_addr, buf, len / 4); | |
330 | else | |
331 | outsl(data_addr, buf, len / 4); | |
332 | ||
22cdd6ce | 333 | if ((io_32bit & 2) && !mmio) |
1da177e4 | 334 | local_irq_restore(flags); |
1da177e4 | 335 | |
16bb69c1 BZ |
336 | if ((len & 3) >= 2) { |
337 | if (mmio) | |
338 | __ide_mm_outsw((void __iomem *)data_addr, | |
339 | (u8 *)buf + (len & ~3), 1); | |
340 | else | |
341 | outsw(data_addr, (u8 *)buf + (len & ~3), 1); | |
342 | } | |
343 | } else { | |
344 | if (mmio) | |
345 | __ide_mm_outsw((void __iomem *)data_addr, buf, len / 2); | |
346 | else | |
347 | outsw(data_addr, buf, len / 2); | |
348 | } | |
1da177e4 LT |
349 | } |
350 | ||
351 | void default_hwif_transport(ide_hwif_t *hwif) | |
352 | { | |
c6dfa867 | 353 | hwif->exec_command = ide_exec_command; |
b73c7ee2 | 354 | hwif->read_status = ide_read_status; |
1f6d8a0f | 355 | hwif->read_altstatus = ide_read_altstatus; |
b2f951aa BZ |
356 | hwif->read_sff_dma_status = ide_read_sff_dma_status; |
357 | ||
6e6afb3b BZ |
358 | hwif->set_irq = ide_set_irq; |
359 | ||
94cd5b62 BZ |
360 | hwif->tf_load = ide_tf_load; |
361 | hwif->tf_read = ide_tf_read; | |
362 | ||
9567b349 BZ |
363 | hwif->input_data = ata_input_data; |
364 | hwif->output_data = ata_output_data; | |
1da177e4 LT |
365 | } |
366 | ||
92eb4380 BZ |
367 | u8 ide_read_error(ide_drive_t *drive) |
368 | { | |
369 | ide_task_t task; | |
370 | ||
371 | memset(&task, 0, sizeof(task)); | |
372 | task.tf_flags = IDE_TFLAG_IN_FEATURE; | |
373 | ||
374 | drive->hwif->tf_read(drive, &task); | |
375 | ||
376 | return task.tf.error; | |
377 | } | |
378 | EXPORT_SYMBOL_GPL(ide_read_error); | |
379 | ||
1823649b BZ |
380 | void ide_read_bcount_and_ireason(ide_drive_t *drive, u16 *bcount, u8 *ireason) |
381 | { | |
382 | ide_task_t task; | |
383 | ||
384 | memset(&task, 0, sizeof(task)); | |
385 | task.tf_flags = IDE_TFLAG_IN_LBAH | IDE_TFLAG_IN_LBAM | | |
386 | IDE_TFLAG_IN_NSECT; | |
387 | ||
388 | drive->hwif->tf_read(drive, &task); | |
389 | ||
390 | *bcount = (task.tf.lbah << 8) | task.tf.lbam; | |
391 | *ireason = task.tf.nsect & 3; | |
392 | } | |
393 | EXPORT_SYMBOL_GPL(ide_read_bcount_and_ireason); | |
394 | ||
1da177e4 LT |
395 | void ide_fix_driveid (struct hd_driveid *id) |
396 | { | |
397 | #ifndef __LITTLE_ENDIAN | |
398 | # ifdef __BIG_ENDIAN | |
399 | int i; | |
400 | u16 *stringcast; | |
401 | ||
402 | id->config = __le16_to_cpu(id->config); | |
403 | id->cyls = __le16_to_cpu(id->cyls); | |
404 | id->reserved2 = __le16_to_cpu(id->reserved2); | |
405 | id->heads = __le16_to_cpu(id->heads); | |
406 | id->track_bytes = __le16_to_cpu(id->track_bytes); | |
407 | id->sector_bytes = __le16_to_cpu(id->sector_bytes); | |
408 | id->sectors = __le16_to_cpu(id->sectors); | |
409 | id->vendor0 = __le16_to_cpu(id->vendor0); | |
410 | id->vendor1 = __le16_to_cpu(id->vendor1); | |
411 | id->vendor2 = __le16_to_cpu(id->vendor2); | |
412 | stringcast = (u16 *)&id->serial_no[0]; | |
413 | for (i = 0; i < (20/2); i++) | |
414 | stringcast[i] = __le16_to_cpu(stringcast[i]); | |
415 | id->buf_type = __le16_to_cpu(id->buf_type); | |
416 | id->buf_size = __le16_to_cpu(id->buf_size); | |
417 | id->ecc_bytes = __le16_to_cpu(id->ecc_bytes); | |
418 | stringcast = (u16 *)&id->fw_rev[0]; | |
419 | for (i = 0; i < (8/2); i++) | |
420 | stringcast[i] = __le16_to_cpu(stringcast[i]); | |
421 | stringcast = (u16 *)&id->model[0]; | |
422 | for (i = 0; i < (40/2); i++) | |
423 | stringcast[i] = __le16_to_cpu(stringcast[i]); | |
424 | id->dword_io = __le16_to_cpu(id->dword_io); | |
425 | id->reserved50 = __le16_to_cpu(id->reserved50); | |
426 | id->field_valid = __le16_to_cpu(id->field_valid); | |
427 | id->cur_cyls = __le16_to_cpu(id->cur_cyls); | |
428 | id->cur_heads = __le16_to_cpu(id->cur_heads); | |
429 | id->cur_sectors = __le16_to_cpu(id->cur_sectors); | |
430 | id->cur_capacity0 = __le16_to_cpu(id->cur_capacity0); | |
431 | id->cur_capacity1 = __le16_to_cpu(id->cur_capacity1); | |
432 | id->lba_capacity = __le32_to_cpu(id->lba_capacity); | |
433 | id->dma_1word = __le16_to_cpu(id->dma_1word); | |
434 | id->dma_mword = __le16_to_cpu(id->dma_mword); | |
435 | id->eide_pio_modes = __le16_to_cpu(id->eide_pio_modes); | |
436 | id->eide_dma_min = __le16_to_cpu(id->eide_dma_min); | |
437 | id->eide_dma_time = __le16_to_cpu(id->eide_dma_time); | |
438 | id->eide_pio = __le16_to_cpu(id->eide_pio); | |
439 | id->eide_pio_iordy = __le16_to_cpu(id->eide_pio_iordy); | |
440 | for (i = 0; i < 2; ++i) | |
441 | id->words69_70[i] = __le16_to_cpu(id->words69_70[i]); | |
442 | for (i = 0; i < 4; ++i) | |
443 | id->words71_74[i] = __le16_to_cpu(id->words71_74[i]); | |
444 | id->queue_depth = __le16_to_cpu(id->queue_depth); | |
445 | for (i = 0; i < 4; ++i) | |
446 | id->words76_79[i] = __le16_to_cpu(id->words76_79[i]); | |
447 | id->major_rev_num = __le16_to_cpu(id->major_rev_num); | |
448 | id->minor_rev_num = __le16_to_cpu(id->minor_rev_num); | |
449 | id->command_set_1 = __le16_to_cpu(id->command_set_1); | |
450 | id->command_set_2 = __le16_to_cpu(id->command_set_2); | |
451 | id->cfsse = __le16_to_cpu(id->cfsse); | |
452 | id->cfs_enable_1 = __le16_to_cpu(id->cfs_enable_1); | |
453 | id->cfs_enable_2 = __le16_to_cpu(id->cfs_enable_2); | |
454 | id->csf_default = __le16_to_cpu(id->csf_default); | |
455 | id->dma_ultra = __le16_to_cpu(id->dma_ultra); | |
456 | id->trseuc = __le16_to_cpu(id->trseuc); | |
457 | id->trsEuc = __le16_to_cpu(id->trsEuc); | |
458 | id->CurAPMvalues = __le16_to_cpu(id->CurAPMvalues); | |
459 | id->mprc = __le16_to_cpu(id->mprc); | |
460 | id->hw_config = __le16_to_cpu(id->hw_config); | |
461 | id->acoustic = __le16_to_cpu(id->acoustic); | |
462 | id->msrqs = __le16_to_cpu(id->msrqs); | |
463 | id->sxfert = __le16_to_cpu(id->sxfert); | |
464 | id->sal = __le16_to_cpu(id->sal); | |
465 | id->spg = __le32_to_cpu(id->spg); | |
466 | id->lba_capacity_2 = __le64_to_cpu(id->lba_capacity_2); | |
467 | for (i = 0; i < 22; i++) | |
468 | id->words104_125[i] = __le16_to_cpu(id->words104_125[i]); | |
469 | id->last_lun = __le16_to_cpu(id->last_lun); | |
470 | id->word127 = __le16_to_cpu(id->word127); | |
471 | id->dlf = __le16_to_cpu(id->dlf); | |
472 | id->csfo = __le16_to_cpu(id->csfo); | |
473 | for (i = 0; i < 26; i++) | |
474 | id->words130_155[i] = __le16_to_cpu(id->words130_155[i]); | |
475 | id->word156 = __le16_to_cpu(id->word156); | |
476 | for (i = 0; i < 3; i++) | |
477 | id->words157_159[i] = __le16_to_cpu(id->words157_159[i]); | |
478 | id->cfa_power = __le16_to_cpu(id->cfa_power); | |
479 | for (i = 0; i < 14; i++) | |
480 | id->words161_175[i] = __le16_to_cpu(id->words161_175[i]); | |
481 | for (i = 0; i < 31; i++) | |
482 | id->words176_205[i] = __le16_to_cpu(id->words176_205[i]); | |
483 | for (i = 0; i < 48; i++) | |
484 | id->words206_254[i] = __le16_to_cpu(id->words206_254[i]); | |
485 | id->integrity_word = __le16_to_cpu(id->integrity_word); | |
486 | # else | |
487 | # error "Please fix <asm/byteorder.h>" | |
488 | # endif | |
489 | #endif | |
490 | } | |
491 | ||
01745112 BZ |
492 | /* |
493 | * ide_fixstring() cleans up and (optionally) byte-swaps a text string, | |
494 | * removing leading/trailing blanks and compressing internal blanks. | |
495 | * It is primarily used to tidy up the model name/number fields as | |
496 | * returned by the WIN_[P]IDENTIFY commands. | |
497 | */ | |
498 | ||
1da177e4 LT |
499 | void ide_fixstring (u8 *s, const int bytecount, const int byteswap) |
500 | { | |
501 | u8 *p = s, *end = &s[bytecount & ~1]; /* bytecount must be even */ | |
502 | ||
503 | if (byteswap) { | |
504 | /* convert from big-endian to host byte order */ | |
505 | for (p = end ; p != s;) { | |
506 | unsigned short *pp = (unsigned short *) (p -= 2); | |
507 | *pp = ntohs(*pp); | |
508 | } | |
509 | } | |
510 | /* strip leading blanks */ | |
511 | while (s != end && *s == ' ') | |
512 | ++s; | |
513 | /* compress internal blanks and strip trailing blanks */ | |
514 | while (s != end && *s) { | |
515 | if (*s++ != ' ' || (s != end && *s && *s != ' ')) | |
516 | *p++ = *(s-1); | |
517 | } | |
518 | /* wipe out trailing garbage */ | |
519 | while (p != end) | |
520 | *p++ = '\0'; | |
521 | } | |
522 | ||
523 | EXPORT_SYMBOL(ide_fixstring); | |
524 | ||
525 | /* | |
526 | * Needed for PCI irq sharing | |
527 | */ | |
528 | int drive_is_ready (ide_drive_t *drive) | |
529 | { | |
530 | ide_hwif_t *hwif = HWIF(drive); | |
531 | u8 stat = 0; | |
532 | ||
533 | if (drive->waiting_for_dma) | |
5e37bdc0 | 534 | return hwif->dma_ops->dma_test_irq(drive); |
1da177e4 LT |
535 | |
536 | #if 0 | |
537 | /* need to guarantee 400ns since last command was issued */ | |
538 | udelay(1); | |
539 | #endif | |
540 | ||
1da177e4 LT |
541 | /* |
542 | * We do a passive status test under shared PCI interrupts on | |
543 | * cards that truly share the ATA side interrupt, but may also share | |
544 | * an interrupt with another pci card/device. We make no assumptions | |
545 | * about possible isa-pnp and pci-pnp issues yet. | |
546 | */ | |
4c3032d8 | 547 | if (hwif->io_ports.ctl_addr) |
1f6d8a0f | 548 | stat = hwif->read_altstatus(hwif); |
1da177e4 | 549 | else |
1da177e4 | 550 | /* Note: this may clear a pending IRQ!! */ |
b73c7ee2 | 551 | stat = hwif->read_status(hwif); |
1da177e4 LT |
552 | |
553 | if (stat & BUSY_STAT) | |
554 | /* drive busy: definitely not interrupting */ | |
555 | return 0; | |
556 | ||
557 | /* drive ready: *might* be interrupting */ | |
558 | return 1; | |
559 | } | |
560 | ||
561 | EXPORT_SYMBOL(drive_is_ready); | |
562 | ||
1da177e4 LT |
563 | /* |
564 | * This routine busy-waits for the drive status to be not "busy". | |
565 | * It then checks the status for all of the "good" bits and none | |
566 | * of the "bad" bits, and if all is okay it returns 0. All other | |
74af21cf | 567 | * cases return error -- caller may then invoke ide_error(). |
1da177e4 LT |
568 | * |
569 | * This routine should get fixed to not hog the cpu during extra long waits.. | |
570 | * That could be done by busy-waiting for the first jiffy or two, and then | |
571 | * setting a timer to wake up at half second intervals thereafter, | |
572 | * until timeout is achieved, before timing out. | |
573 | */ | |
aedea591 | 574 | static int __ide_wait_stat(ide_drive_t *drive, u8 good, u8 bad, unsigned long timeout, u8 *rstat) |
1da177e4 | 575 | { |
b73c7ee2 | 576 | ide_hwif_t *hwif = drive->hwif; |
1da177e4 | 577 | unsigned long flags; |
74af21cf BZ |
578 | int i; |
579 | u8 stat; | |
1da177e4 LT |
580 | |
581 | udelay(1); /* spec allows drive 400ns to assert "BUSY" */ | |
b73c7ee2 | 582 | stat = hwif->read_status(hwif); |
c47137a9 BZ |
583 | |
584 | if (stat & BUSY_STAT) { | |
1da177e4 LT |
585 | local_irq_set(flags); |
586 | timeout += jiffies; | |
b73c7ee2 | 587 | while ((stat = hwif->read_status(hwif)) & BUSY_STAT) { |
1da177e4 LT |
588 | if (time_after(jiffies, timeout)) { |
589 | /* | |
590 | * One last read after the timeout in case | |
591 | * heavy interrupt load made us not make any | |
592 | * progress during the timeout.. | |
593 | */ | |
b73c7ee2 | 594 | stat = hwif->read_status(hwif); |
1da177e4 LT |
595 | if (!(stat & BUSY_STAT)) |
596 | break; | |
597 | ||
598 | local_irq_restore(flags); | |
74af21cf BZ |
599 | *rstat = stat; |
600 | return -EBUSY; | |
1da177e4 LT |
601 | } |
602 | } | |
603 | local_irq_restore(flags); | |
604 | } | |
605 | /* | |
606 | * Allow status to settle, then read it again. | |
607 | * A few rare drives vastly violate the 400ns spec here, | |
608 | * so we'll wait up to 10usec for a "good" status | |
609 | * rather than expensively fail things immediately. | |
610 | * This fix courtesy of Matthew Faupel & Niccolo Rigacci. | |
611 | */ | |
612 | for (i = 0; i < 10; i++) { | |
613 | udelay(1); | |
b73c7ee2 | 614 | stat = hwif->read_status(hwif); |
c47137a9 BZ |
615 | |
616 | if (OK_STAT(stat, good, bad)) { | |
74af21cf | 617 | *rstat = stat; |
1da177e4 | 618 | return 0; |
74af21cf | 619 | } |
1da177e4 | 620 | } |
74af21cf BZ |
621 | *rstat = stat; |
622 | return -EFAULT; | |
623 | } | |
624 | ||
625 | /* | |
626 | * In case of error returns error value after doing "*startstop = ide_error()". | |
627 | * The caller should return the updated value of "startstop" in this case, | |
628 | * "startstop" is unchanged when the function returns 0. | |
629 | */ | |
630 | int ide_wait_stat(ide_startstop_t *startstop, ide_drive_t *drive, u8 good, u8 bad, unsigned long timeout) | |
631 | { | |
632 | int err; | |
633 | u8 stat; | |
634 | ||
635 | /* bail early if we've exceeded max_failures */ | |
636 | if (drive->max_failures && (drive->failures > drive->max_failures)) { | |
637 | *startstop = ide_stopped; | |
638 | return 1; | |
639 | } | |
640 | ||
641 | err = __ide_wait_stat(drive, good, bad, timeout, &stat); | |
642 | ||
643 | if (err) { | |
644 | char *s = (err == -EBUSY) ? "status timeout" : "status error"; | |
645 | *startstop = ide_error(drive, s, stat); | |
646 | } | |
647 | ||
648 | return err; | |
1da177e4 LT |
649 | } |
650 | ||
651 | EXPORT_SYMBOL(ide_wait_stat); | |
652 | ||
a5b7e70d BZ |
653 | /** |
654 | * ide_in_drive_list - look for drive in black/white list | |
655 | * @id: drive identifier | |
656 | * @drive_table: list to inspect | |
657 | * | |
658 | * Look for a drive in the blacklist and the whitelist tables | |
659 | * Returns 1 if the drive is found in the table. | |
660 | */ | |
661 | ||
662 | int ide_in_drive_list(struct hd_driveid *id, const struct drive_list_entry *drive_table) | |
663 | { | |
664 | for ( ; drive_table->id_model; drive_table++) | |
665 | if ((!strcmp(drive_table->id_model, id->model)) && | |
666 | (!drive_table->id_firmware || | |
667 | strstr(id->fw_rev, drive_table->id_firmware))) | |
668 | return 1; | |
669 | return 0; | |
670 | } | |
671 | ||
b0244a00 BZ |
672 | EXPORT_SYMBOL_GPL(ide_in_drive_list); |
673 | ||
a5b7e70d BZ |
674 | /* |
675 | * Early UDMA66 devices don't set bit14 to 1, only bit13 is valid. | |
676 | * We list them here and depend on the device side cable detection for them. | |
8588a2b7 BZ |
677 | * |
678 | * Some optical devices with the buggy firmwares have the same problem. | |
a5b7e70d BZ |
679 | */ |
680 | static const struct drive_list_entry ivb_list[] = { | |
681 | { "QUANTUM FIREBALLlct10 05" , "A03.0900" }, | |
8588a2b7 | 682 | { "TSSTcorp CDDVDW SH-S202J" , "SB00" }, |
e97564f3 PM |
683 | { "TSSTcorp CDDVDW SH-S202J" , "SB01" }, |
684 | { "TSSTcorp CDDVDW SH-S202N" , "SB00" }, | |
685 | { "TSSTcorp CDDVDW SH-S202N" , "SB01" }, | |
3ced5c49 AS |
686 | { "TSSTcorp CDDVDW SH-S202H" , "SB00" }, |
687 | { "TSSTcorp CDDVDW SH-S202H" , "SB01" }, | |
a5b7e70d BZ |
688 | { NULL , NULL } |
689 | }; | |
690 | ||
1da177e4 LT |
691 | /* |
692 | * All hosts that use the 80c ribbon must use! | |
693 | * The name is derived from upper byte of word 93 and the 80c ribbon. | |
694 | */ | |
695 | u8 eighty_ninty_three (ide_drive_t *drive) | |
696 | { | |
7f8f48af BZ |
697 | ide_hwif_t *hwif = drive->hwif; |
698 | struct hd_driveid *id = drive->id; | |
a5b7e70d | 699 | int ivb = ide_in_drive_list(id, ivb_list); |
7f8f48af | 700 | |
49521f97 BZ |
701 | if (hwif->cbl == ATA_CBL_PATA40_SHORT) |
702 | return 1; | |
703 | ||
a5b7e70d BZ |
704 | if (ivb) |
705 | printk(KERN_DEBUG "%s: skipping word 93 validity check\n", | |
706 | drive->name); | |
707 | ||
b98f8803 GK |
708 | if (ide_dev_is_sata(id) && !ivb) |
709 | return 1; | |
710 | ||
a5b7e70d | 711 | if (hwif->cbl != ATA_CBL_PATA80 && !ivb) |
7f8f48af | 712 | goto no_80w; |
1a1276e7 | 713 | |
f68d9320 BZ |
714 | /* |
715 | * FIXME: | |
f367bed0 | 716 | * - change master/slave IDENTIFY order |
a5b7e70d | 717 | * - force bit13 (80c cable present) check also for !ivb devices |
f68d9320 BZ |
718 | * (unless the slave device is pre-ATA3) |
719 | */ | |
a5b7e70d | 720 | if ((id->hw_config & 0x4000) || (ivb && (id->hw_config & 0x2000))) |
7f8f48af BZ |
721 | return 1; |
722 | ||
723 | no_80w: | |
724 | if (drive->udma33_warned == 1) | |
725 | return 0; | |
726 | ||
727 | printk(KERN_WARNING "%s: %s side 80-wire cable detection failed, " | |
728 | "limiting max speed to UDMA33\n", | |
49521f97 BZ |
729 | drive->name, |
730 | hwif->cbl == ATA_CBL_PATA80 ? "drive" : "host"); | |
7f8f48af BZ |
731 | |
732 | drive->udma33_warned = 1; | |
733 | ||
734 | return 0; | |
1da177e4 LT |
735 | } |
736 | ||
8a455134 | 737 | int ide_driveid_update(ide_drive_t *drive) |
1da177e4 | 738 | { |
8a455134 | 739 | ide_hwif_t *hwif = drive->hwif; |
1da177e4 | 740 | struct hd_driveid *id; |
8a455134 | 741 | unsigned long timeout, flags; |
c47137a9 | 742 | u8 stat; |
1da177e4 | 743 | |
1da177e4 LT |
744 | /* |
745 | * Re-read drive->id for possible DMA mode | |
746 | * change (copied from ide-probe.c) | |
747 | */ | |
1da177e4 LT |
748 | |
749 | SELECT_MASK(drive, 1); | |
6e6afb3b | 750 | hwif->set_irq(hwif, 0); |
1da177e4 | 751 | msleep(50); |
c6dfa867 | 752 | hwif->exec_command(hwif, WIN_IDENTIFY); |
1da177e4 LT |
753 | timeout = jiffies + WAIT_WORSTCASE; |
754 | do { | |
755 | if (time_after(jiffies, timeout)) { | |
756 | SELECT_MASK(drive, 0); | |
757 | return 0; /* drive timed-out */ | |
758 | } | |
c47137a9 | 759 | |
1da177e4 | 760 | msleep(50); /* give drive a breather */ |
1f6d8a0f | 761 | stat = hwif->read_altstatus(hwif); |
c47137a9 BZ |
762 | } while (stat & BUSY_STAT); |
763 | ||
1da177e4 | 764 | msleep(50); /* wait for IRQ and DRQ_STAT */ |
b73c7ee2 | 765 | stat = hwif->read_status(hwif); |
c47137a9 BZ |
766 | |
767 | if (!OK_STAT(stat, DRQ_STAT, BAD_R_STAT)) { | |
1da177e4 LT |
768 | SELECT_MASK(drive, 0); |
769 | printk("%s: CHECK for good STATUS\n", drive->name); | |
770 | return 0; | |
771 | } | |
772 | local_irq_save(flags); | |
773 | SELECT_MASK(drive, 0); | |
774 | id = kmalloc(SECTOR_WORDS*4, GFP_ATOMIC); | |
775 | if (!id) { | |
776 | local_irq_restore(flags); | |
777 | return 0; | |
778 | } | |
9567b349 | 779 | hwif->input_data(drive, NULL, id, SECTOR_SIZE); |
b73c7ee2 | 780 | (void)hwif->read_status(hwif); /* clear drive IRQ */ |
1da177e4 LT |
781 | local_irq_enable(); |
782 | local_irq_restore(flags); | |
783 | ide_fix_driveid(id); | |
784 | if (id) { | |
785 | drive->id->dma_ultra = id->dma_ultra; | |
786 | drive->id->dma_mword = id->dma_mword; | |
787 | drive->id->dma_1word = id->dma_1word; | |
788 | /* anything more ? */ | |
789 | kfree(id); | |
3ab7efe8 BZ |
790 | |
791 | if (drive->using_dma && ide_id_dma_bug(drive)) | |
792 | ide_dma_off(drive); | |
1da177e4 LT |
793 | } |
794 | ||
795 | return 1; | |
1da177e4 LT |
796 | } |
797 | ||
74af21cf | 798 | int ide_config_drive_speed(ide_drive_t *drive, u8 speed) |
1da177e4 | 799 | { |
74af21cf | 800 | ide_hwif_t *hwif = drive->hwif; |
89613e66 | 801 | int error = 0; |
1da177e4 | 802 | u8 stat; |
59be2c80 | 803 | ide_task_t task; |
1da177e4 | 804 | |
1da177e4 | 805 | #ifdef CONFIG_BLK_DEV_IDEDMA |
5e37bdc0 BZ |
806 | if (hwif->dma_ops) /* check if host supports DMA */ |
807 | hwif->dma_ops->dma_host_set(drive, 0); | |
1da177e4 LT |
808 | #endif |
809 | ||
89613e66 SS |
810 | /* Skip setting PIO flow-control modes on pre-EIDE drives */ |
811 | if ((speed & 0xf8) == XFER_PIO_0 && !(drive->id->capability & 0x08)) | |
812 | goto skip; | |
813 | ||
1da177e4 LT |
814 | /* |
815 | * Don't use ide_wait_cmd here - it will | |
816 | * attempt to set_geometry and recalibrate, | |
817 | * but for some reason these don't work at | |
818 | * this point (lost interrupt). | |
819 | */ | |
820 | /* | |
821 | * Select the drive, and issue the SETFEATURES command | |
822 | */ | |
823 | disable_irq_nosync(hwif->irq); | |
824 | ||
825 | /* | |
826 | * FIXME: we race against the running IRQ here if | |
827 | * this is called from non IRQ context. If we use | |
828 | * disable_irq() we hang on the error path. Work | |
829 | * is needed. | |
830 | */ | |
831 | ||
832 | udelay(1); | |
833 | SELECT_DRIVE(drive); | |
834 | SELECT_MASK(drive, 0); | |
835 | udelay(1); | |
6e6afb3b | 836 | hwif->set_irq(hwif, 0); |
59be2c80 BZ |
837 | |
838 | memset(&task, 0, sizeof(task)); | |
839 | task.tf_flags = IDE_TFLAG_OUT_FEATURE | IDE_TFLAG_OUT_NSECT; | |
840 | task.tf.feature = SETFEATURES_XFER; | |
841 | task.tf.nsect = speed; | |
842 | ||
843 | hwif->tf_load(drive, &task); | |
844 | ||
c6dfa867 | 845 | hwif->exec_command(hwif, WIN_SETFEATURES); |
59be2c80 | 846 | |
81ca6919 | 847 | if (drive->quirk_list == 2) |
6e6afb3b | 848 | hwif->set_irq(hwif, 1); |
1da177e4 | 849 | |
74af21cf BZ |
850 | error = __ide_wait_stat(drive, drive->ready_stat, |
851 | BUSY_STAT|DRQ_STAT|ERR_STAT, | |
852 | WAIT_CMD, &stat); | |
1da177e4 LT |
853 | |
854 | SELECT_MASK(drive, 0); | |
855 | ||
856 | enable_irq(hwif->irq); | |
857 | ||
858 | if (error) { | |
859 | (void) ide_dump_status(drive, "set_drive_speed_status", stat); | |
860 | return error; | |
861 | } | |
862 | ||
863 | drive->id->dma_ultra &= ~0xFF00; | |
864 | drive->id->dma_mword &= ~0x0F00; | |
865 | drive->id->dma_1word &= ~0x0F00; | |
866 | ||
89613e66 | 867 | skip: |
1da177e4 | 868 | #ifdef CONFIG_BLK_DEV_IDEDMA |
ba4b2e60 | 869 | if (speed >= XFER_SW_DMA_0 && drive->using_dma) |
5e37bdc0 BZ |
870 | hwif->dma_ops->dma_host_set(drive, 1); |
871 | else if (hwif->dma_ops) /* check if host supports DMA */ | |
4a546e04 | 872 | ide_dma_off_quietly(drive); |
1da177e4 LT |
873 | #endif |
874 | ||
875 | switch(speed) { | |
876 | case XFER_UDMA_7: drive->id->dma_ultra |= 0x8080; break; | |
877 | case XFER_UDMA_6: drive->id->dma_ultra |= 0x4040; break; | |
878 | case XFER_UDMA_5: drive->id->dma_ultra |= 0x2020; break; | |
879 | case XFER_UDMA_4: drive->id->dma_ultra |= 0x1010; break; | |
880 | case XFER_UDMA_3: drive->id->dma_ultra |= 0x0808; break; | |
881 | case XFER_UDMA_2: drive->id->dma_ultra |= 0x0404; break; | |
882 | case XFER_UDMA_1: drive->id->dma_ultra |= 0x0202; break; | |
883 | case XFER_UDMA_0: drive->id->dma_ultra |= 0x0101; break; | |
884 | case XFER_MW_DMA_2: drive->id->dma_mword |= 0x0404; break; | |
885 | case XFER_MW_DMA_1: drive->id->dma_mword |= 0x0202; break; | |
886 | case XFER_MW_DMA_0: drive->id->dma_mword |= 0x0101; break; | |
887 | case XFER_SW_DMA_2: drive->id->dma_1word |= 0x0404; break; | |
888 | case XFER_SW_DMA_1: drive->id->dma_1word |= 0x0202; break; | |
889 | case XFER_SW_DMA_0: drive->id->dma_1word |= 0x0101; break; | |
890 | default: break; | |
891 | } | |
892 | if (!drive->init_speed) | |
893 | drive->init_speed = speed; | |
894 | drive->current_speed = speed; | |
895 | return error; | |
896 | } | |
897 | ||
1da177e4 LT |
898 | /* |
899 | * This should get invoked any time we exit the driver to | |
900 | * wait for an interrupt response from a drive. handler() points | |
901 | * at the appropriate code to handle the next interrupt, and a | |
902 | * timer is started to prevent us from waiting forever in case | |
903 | * something goes wrong (see the ide_timer_expiry() handler later on). | |
904 | * | |
905 | * See also ide_execute_command | |
906 | */ | |
907 | static void __ide_set_handler (ide_drive_t *drive, ide_handler_t *handler, | |
908 | unsigned int timeout, ide_expiry_t *expiry) | |
909 | { | |
910 | ide_hwgroup_t *hwgroup = HWGROUP(drive); | |
911 | ||
d30a426d | 912 | BUG_ON(hwgroup->handler); |
1da177e4 LT |
913 | hwgroup->handler = handler; |
914 | hwgroup->expiry = expiry; | |
915 | hwgroup->timer.expires = jiffies + timeout; | |
d30a426d | 916 | hwgroup->req_gen_timer = hwgroup->req_gen; |
1da177e4 LT |
917 | add_timer(&hwgroup->timer); |
918 | } | |
919 | ||
920 | void ide_set_handler (ide_drive_t *drive, ide_handler_t *handler, | |
921 | unsigned int timeout, ide_expiry_t *expiry) | |
922 | { | |
923 | unsigned long flags; | |
924 | spin_lock_irqsave(&ide_lock, flags); | |
925 | __ide_set_handler(drive, handler, timeout, expiry); | |
926 | spin_unlock_irqrestore(&ide_lock, flags); | |
927 | } | |
928 | ||
929 | EXPORT_SYMBOL(ide_set_handler); | |
930 | ||
931 | /** | |
932 | * ide_execute_command - execute an IDE command | |
933 | * @drive: IDE drive to issue the command against | |
934 | * @command: command byte to write | |
935 | * @handler: handler for next phase | |
936 | * @timeout: timeout for command | |
937 | * @expiry: handler to run on timeout | |
938 | * | |
939 | * Helper function to issue an IDE command. This handles the | |
940 | * atomicity requirements, command timing and ensures that the | |
941 | * handler and IRQ setup do not race. All IDE command kick off | |
942 | * should go via this function or do equivalent locking. | |
943 | */ | |
cd2a2d96 BZ |
944 | |
945 | void ide_execute_command(ide_drive_t *drive, u8 cmd, ide_handler_t *handler, | |
946 | unsigned timeout, ide_expiry_t *expiry) | |
1da177e4 LT |
947 | { |
948 | unsigned long flags; | |
1da177e4 | 949 | ide_hwif_t *hwif = HWIF(drive); |
629f944b | 950 | |
1da177e4 | 951 | spin_lock_irqsave(&ide_lock, flags); |
629f944b | 952 | __ide_set_handler(drive, handler, timeout, expiry); |
c6dfa867 | 953 | hwif->exec_command(hwif, cmd); |
629f944b BZ |
954 | /* |
955 | * Drive takes 400nS to respond, we must avoid the IRQ being | |
956 | * serviced before that. | |
957 | * | |
958 | * FIXME: we could skip this delay with care on non shared devices | |
959 | */ | |
1da177e4 LT |
960 | ndelay(400); |
961 | spin_unlock_irqrestore(&ide_lock, flags); | |
962 | } | |
1da177e4 LT |
963 | EXPORT_SYMBOL(ide_execute_command); |
964 | ||
1fc14258 BZ |
965 | void ide_execute_pkt_cmd(ide_drive_t *drive) |
966 | { | |
967 | ide_hwif_t *hwif = drive->hwif; | |
968 | unsigned long flags; | |
969 | ||
970 | spin_lock_irqsave(&ide_lock, flags); | |
c6dfa867 | 971 | hwif->exec_command(hwif, WIN_PACKETCMD); |
1fc14258 BZ |
972 | ndelay(400); |
973 | spin_unlock_irqrestore(&ide_lock, flags); | |
974 | } | |
975 | EXPORT_SYMBOL_GPL(ide_execute_pkt_cmd); | |
1da177e4 | 976 | |
64a8f00f | 977 | static inline void ide_complete_drive_reset(ide_drive_t *drive, int err) |
79e36a9f EO |
978 | { |
979 | struct request *rq = drive->hwif->hwgroup->rq; | |
980 | ||
981 | if (rq && blk_special_request(rq) && rq->cmd[0] == REQ_DRIVE_RESET) | |
64a8f00f | 982 | ide_end_request(drive, err ? err : 1, 0); |
79e36a9f EO |
983 | } |
984 | ||
1da177e4 LT |
985 | /* needed below */ |
986 | static ide_startstop_t do_reset1 (ide_drive_t *, int); | |
987 | ||
988 | /* | |
989 | * atapi_reset_pollfunc() gets invoked to poll the interface for completion every 50ms | |
990 | * during an atapi drive reset operation. If the drive has not yet responded, | |
991 | * and we have not yet hit our maximum waiting time, then the timer is restarted | |
992 | * for another 50ms. | |
993 | */ | |
994 | static ide_startstop_t atapi_reset_pollfunc (ide_drive_t *drive) | |
995 | { | |
b73c7ee2 BZ |
996 | ide_hwif_t *hwif = drive->hwif; |
997 | ide_hwgroup_t *hwgroup = hwif->hwgroup; | |
1da177e4 LT |
998 | u8 stat; |
999 | ||
1000 | SELECT_DRIVE(drive); | |
1001 | udelay (10); | |
b73c7ee2 | 1002 | stat = hwif->read_status(hwif); |
1da177e4 | 1003 | |
c47137a9 | 1004 | if (OK_STAT(stat, 0, BUSY_STAT)) |
1da177e4 | 1005 | printk("%s: ATAPI reset complete\n", drive->name); |
c47137a9 | 1006 | else { |
1da177e4 | 1007 | if (time_before(jiffies, hwgroup->poll_timeout)) { |
1da177e4 LT |
1008 | ide_set_handler(drive, &atapi_reset_pollfunc, HZ/20, NULL); |
1009 | /* continue polling */ | |
1010 | return ide_started; | |
1011 | } | |
1012 | /* end of polling */ | |
1013 | hwgroup->polling = 0; | |
1014 | printk("%s: ATAPI reset timed-out, status=0x%02x\n", | |
1015 | drive->name, stat); | |
1016 | /* do it the old fashioned way */ | |
1017 | return do_reset1(drive, 1); | |
1018 | } | |
1019 | /* done polling */ | |
1020 | hwgroup->polling = 0; | |
64a8f00f | 1021 | ide_complete_drive_reset(drive, 0); |
1da177e4 LT |
1022 | return ide_stopped; |
1023 | } | |
1024 | ||
1025 | /* | |
1026 | * reset_pollfunc() gets invoked to poll the interface for completion every 50ms | |
1027 | * during an ide reset operation. If the drives have not yet responded, | |
1028 | * and we have not yet hit our maximum waiting time, then the timer is restarted | |
1029 | * for another 50ms. | |
1030 | */ | |
1031 | static ide_startstop_t reset_pollfunc (ide_drive_t *drive) | |
1032 | { | |
1033 | ide_hwgroup_t *hwgroup = HWGROUP(drive); | |
1034 | ide_hwif_t *hwif = HWIF(drive); | |
ac95beed | 1035 | const struct ide_port_ops *port_ops = hwif->port_ops; |
1da177e4 | 1036 | u8 tmp; |
64a8f00f | 1037 | int err = 0; |
1da177e4 | 1038 | |
ac95beed | 1039 | if (port_ops && port_ops->reset_poll) { |
64a8f00f EO |
1040 | err = port_ops->reset_poll(drive); |
1041 | if (err) { | |
1da177e4 LT |
1042 | printk(KERN_ERR "%s: host reset_poll failure for %s.\n", |
1043 | hwif->name, drive->name); | |
79e36a9f | 1044 | goto out; |
1da177e4 LT |
1045 | } |
1046 | } | |
1047 | ||
b73c7ee2 | 1048 | tmp = hwif->read_status(hwif); |
c47137a9 BZ |
1049 | |
1050 | if (!OK_STAT(tmp, 0, BUSY_STAT)) { | |
1da177e4 | 1051 | if (time_before(jiffies, hwgroup->poll_timeout)) { |
1da177e4 LT |
1052 | ide_set_handler(drive, &reset_pollfunc, HZ/20, NULL); |
1053 | /* continue polling */ | |
1054 | return ide_started; | |
1055 | } | |
1056 | printk("%s: reset timed-out, status=0x%02x\n", hwif->name, tmp); | |
1057 | drive->failures++; | |
64a8f00f | 1058 | err = -EIO; |
1da177e4 LT |
1059 | } else { |
1060 | printk("%s: reset: ", hwif->name); | |
64a57fe4 BZ |
1061 | tmp = ide_read_error(drive); |
1062 | ||
1063 | if (tmp == 1) { | |
1da177e4 LT |
1064 | printk("success\n"); |
1065 | drive->failures = 0; | |
1066 | } else { | |
1067 | drive->failures++; | |
1068 | printk("master: "); | |
1069 | switch (tmp & 0x7f) { | |
1070 | case 1: printk("passed"); | |
1071 | break; | |
1072 | case 2: printk("formatter device error"); | |
1073 | break; | |
1074 | case 3: printk("sector buffer error"); | |
1075 | break; | |
1076 | case 4: printk("ECC circuitry error"); | |
1077 | break; | |
1078 | case 5: printk("controlling MPU error"); | |
1079 | break; | |
1080 | default:printk("error (0x%02x?)", tmp); | |
1081 | } | |
1082 | if (tmp & 0x80) | |
1083 | printk("; slave: failed"); | |
1084 | printk("\n"); | |
64a8f00f | 1085 | err = -EIO; |
1da177e4 LT |
1086 | } |
1087 | } | |
79e36a9f | 1088 | out: |
64a8f00f EO |
1089 | hwgroup->polling = 0; /* done polling */ |
1090 | ide_complete_drive_reset(drive, err); | |
1da177e4 LT |
1091 | return ide_stopped; |
1092 | } | |
1093 | ||
1da177e4 LT |
1094 | static void ide_disk_pre_reset(ide_drive_t *drive) |
1095 | { | |
1096 | int legacy = (drive->id->cfs_enable_2 & 0x0400) ? 0 : 1; | |
1097 | ||
1098 | drive->special.all = 0; | |
1099 | drive->special.b.set_geometry = legacy; | |
1100 | drive->special.b.recalibrate = legacy; | |
4ee06b7e | 1101 | drive->mult_count = 0; |
1da177e4 LT |
1102 | if (!drive->keep_settings && !drive->using_dma) |
1103 | drive->mult_req = 0; | |
1104 | if (drive->mult_req != drive->mult_count) | |
1105 | drive->special.b.set_multmode = 1; | |
1106 | } | |
1107 | ||
1108 | static void pre_reset(ide_drive_t *drive) | |
1109 | { | |
ac95beed BZ |
1110 | const struct ide_port_ops *port_ops = drive->hwif->port_ops; |
1111 | ||
1da177e4 LT |
1112 | if (drive->media == ide_disk) |
1113 | ide_disk_pre_reset(drive); | |
1114 | else | |
1115 | drive->post_reset = 1; | |
1116 | ||
99ffbe0e BZ |
1117 | if (drive->using_dma) { |
1118 | if (drive->crc_count) | |
578cfa0d | 1119 | ide_check_dma_crc(drive); |
99ffbe0e BZ |
1120 | else |
1121 | ide_dma_off(drive); | |
1122 | } | |
1123 | ||
1124 | if (!drive->keep_settings) { | |
1125 | if (!drive->using_dma) { | |
1da177e4 LT |
1126 | drive->unmask = 0; |
1127 | drive->io_32bit = 0; | |
1128 | } | |
1129 | return; | |
1130 | } | |
1da177e4 | 1131 | |
ac95beed BZ |
1132 | if (port_ops && port_ops->pre_reset) |
1133 | port_ops->pre_reset(drive); | |
1da177e4 | 1134 | |
513daadd SS |
1135 | if (drive->current_speed != 0xff) |
1136 | drive->desired_speed = drive->current_speed; | |
1137 | drive->current_speed = 0xff; | |
1da177e4 LT |
1138 | } |
1139 | ||
1140 | /* | |
1141 | * do_reset1() attempts to recover a confused drive by resetting it. | |
1142 | * Unfortunately, resetting a disk drive actually resets all devices on | |
1143 | * the same interface, so it can really be thought of as resetting the | |
1144 | * interface rather than resetting the drive. | |
1145 | * | |
1146 | * ATAPI devices have their own reset mechanism which allows them to be | |
1147 | * individually reset without clobbering other devices on the same interface. | |
1148 | * | |
1149 | * Unfortunately, the IDE interface does not generate an interrupt to let | |
1150 | * us know when the reset operation has finished, so we must poll for this. | |
1151 | * Equally poor, though, is the fact that this may a very long time to complete, | |
1152 | * (up to 30 seconds worstcase). So, instead of busy-waiting here for it, | |
1153 | * we set a timer to poll at 50ms intervals. | |
1154 | */ | |
1155 | static ide_startstop_t do_reset1 (ide_drive_t *drive, int do_not_try_atapi) | |
1156 | { | |
1157 | unsigned int unit; | |
1158 | unsigned long flags; | |
1159 | ide_hwif_t *hwif; | |
1160 | ide_hwgroup_t *hwgroup; | |
4c3032d8 | 1161 | struct ide_io_ports *io_ports; |
ac95beed | 1162 | const struct ide_port_ops *port_ops; |
23579a2a | 1163 | |
1da177e4 LT |
1164 | spin_lock_irqsave(&ide_lock, flags); |
1165 | hwif = HWIF(drive); | |
1166 | hwgroup = HWGROUP(drive); | |
1167 | ||
4c3032d8 BZ |
1168 | io_ports = &hwif->io_ports; |
1169 | ||
1da177e4 | 1170 | /* We must not reset with running handlers */ |
125e1874 | 1171 | BUG_ON(hwgroup->handler != NULL); |
1da177e4 LT |
1172 | |
1173 | /* For an ATAPI device, first try an ATAPI SRST. */ | |
1174 | if (drive->media != ide_disk && !do_not_try_atapi) { | |
1175 | pre_reset(drive); | |
1176 | SELECT_DRIVE(drive); | |
1177 | udelay (20); | |
c6dfa867 | 1178 | hwif->exec_command(hwif, WIN_SRST); |
68ad9910 | 1179 | ndelay(400); |
1da177e4 LT |
1180 | hwgroup->poll_timeout = jiffies + WAIT_WORSTCASE; |
1181 | hwgroup->polling = 1; | |
1182 | __ide_set_handler(drive, &atapi_reset_pollfunc, HZ/20, NULL); | |
1183 | spin_unlock_irqrestore(&ide_lock, flags); | |
1184 | return ide_started; | |
1185 | } | |
1186 | ||
1187 | /* | |
1188 | * First, reset any device state data we were maintaining | |
1189 | * for any of the drives on this interface. | |
1190 | */ | |
1191 | for (unit = 0; unit < MAX_DRIVES; ++unit) | |
1192 | pre_reset(&hwif->drives[unit]); | |
1193 | ||
4c3032d8 | 1194 | if (io_ports->ctl_addr == 0) { |
1da177e4 | 1195 | spin_unlock_irqrestore(&ide_lock, flags); |
64a8f00f | 1196 | ide_complete_drive_reset(drive, -ENXIO); |
1da177e4 LT |
1197 | return ide_stopped; |
1198 | } | |
1199 | ||
1200 | /* | |
1201 | * Note that we also set nIEN while resetting the device, | |
1202 | * to mask unwanted interrupts from the interface during the reset. | |
1203 | * However, due to the design of PC hardware, this will cause an | |
1204 | * immediate interrupt due to the edge transition it produces. | |
1205 | * This single interrupt gives us a "fast poll" for drives that | |
1206 | * recover from reset very quickly, saving us the first 50ms wait time. | |
6e6afb3b BZ |
1207 | * |
1208 | * TODO: add ->softreset method and stop abusing ->set_irq | |
1da177e4 LT |
1209 | */ |
1210 | /* set SRST and nIEN */ | |
6e6afb3b | 1211 | hwif->set_irq(hwif, 4); |
1da177e4 LT |
1212 | /* more than enough time */ |
1213 | udelay(10); | |
6e6afb3b BZ |
1214 | /* clear SRST, leave nIEN (unless device is on the quirk list) */ |
1215 | hwif->set_irq(hwif, drive->quirk_list == 2); | |
1da177e4 LT |
1216 | /* more than enough time */ |
1217 | udelay(10); | |
1218 | hwgroup->poll_timeout = jiffies + WAIT_WORSTCASE; | |
1219 | hwgroup->polling = 1; | |
1220 | __ide_set_handler(drive, &reset_pollfunc, HZ/20, NULL); | |
1221 | ||
1222 | /* | |
1223 | * Some weird controller like resetting themselves to a strange | |
1224 | * state when the disks are reset this way. At least, the Winbond | |
1225 | * 553 documentation says that | |
1226 | */ | |
ac95beed BZ |
1227 | port_ops = hwif->port_ops; |
1228 | if (port_ops && port_ops->resetproc) | |
1229 | port_ops->resetproc(drive); | |
1da177e4 LT |
1230 | |
1231 | spin_unlock_irqrestore(&ide_lock, flags); | |
1232 | return ide_started; | |
1233 | } | |
1234 | ||
1235 | /* | |
1236 | * ide_do_reset() is the entry point to the drive/interface reset code. | |
1237 | */ | |
1238 | ||
1239 | ide_startstop_t ide_do_reset (ide_drive_t *drive) | |
1240 | { | |
1241 | return do_reset1(drive, 0); | |
1242 | } | |
1243 | ||
1244 | EXPORT_SYMBOL(ide_do_reset); | |
1245 | ||
1246 | /* | |
1247 | * ide_wait_not_busy() waits for the currently selected device on the hwif | |
9d501529 | 1248 | * to report a non-busy status, see comments in ide_probe_port(). |
1da177e4 LT |
1249 | */ |
1250 | int ide_wait_not_busy(ide_hwif_t *hwif, unsigned long timeout) | |
1251 | { | |
1252 | u8 stat = 0; | |
1253 | ||
1254 | while(timeout--) { | |
1255 | /* | |
1256 | * Turn this into a schedule() sleep once I'm sure | |
1257 | * about locking issues (2.5 work ?). | |
1258 | */ | |
1259 | mdelay(1); | |
b73c7ee2 | 1260 | stat = hwif->read_status(hwif); |
1da177e4 LT |
1261 | if ((stat & BUSY_STAT) == 0) |
1262 | return 0; | |
1263 | /* | |
1264 | * Assume a value of 0xff means nothing is connected to | |
1265 | * the interface and it doesn't implement the pull-down | |
1266 | * resistor on D7. | |
1267 | */ | |
1268 | if (stat == 0xff) | |
1269 | return -ENODEV; | |
6842f8c8 | 1270 | touch_softlockup_watchdog(); |
1e86240f | 1271 | touch_nmi_watchdog(); |
1da177e4 LT |
1272 | } |
1273 | return -EBUSY; | |
1274 | } | |
1275 | ||
1276 | EXPORT_SYMBOL_GPL(ide_wait_not_busy); | |
1277 |