]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * libata-scsi.c - helper library for ATA | |
3 | * | |
4 | * Maintained by: Tejun Heo <tj@kernel.org> | |
5 | * Please ALWAYS copy linux-ide@vger.kernel.org | |
6 | * on emails. | |
7 | * | |
8 | * Copyright 2003-2004 Red Hat, Inc. All rights reserved. | |
9 | * Copyright 2003-2004 Jeff Garzik | |
10 | * | |
11 | * | |
12 | * This program is free software; you can redistribute it and/or modify | |
13 | * it under the terms of the GNU General Public License as published by | |
14 | * the Free Software Foundation; either version 2, or (at your option) | |
15 | * any later version. | |
16 | * | |
17 | * This program is distributed in the hope that it will be useful, | |
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
20 | * GNU General Public License for more details. | |
21 | * | |
22 | * You should have received a copy of the GNU General Public License | |
23 | * along with this program; see the file COPYING. If not, write to | |
24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | |
25 | * | |
26 | * | |
27 | * libata documentation is available via 'make {ps|pdf}docs', | |
28 | * as Documentation/DocBook/libata.* | |
29 | * | |
30 | * Hardware documentation available from | |
31 | * - http://www.t10.org/ | |
32 | * - http://www.t13.org/ | |
33 | * | |
34 | */ | |
35 | ||
36 | #include <linux/slab.h> | |
37 | #include <linux/kernel.h> | |
38 | #include <linux/blkdev.h> | |
39 | #include <linux/spinlock.h> | |
40 | #include <linux/export.h> | |
41 | #include <scsi/scsi.h> | |
42 | #include <scsi/scsi_host.h> | |
43 | #include <scsi/scsi_cmnd.h> | |
44 | #include <scsi/scsi_eh.h> | |
45 | #include <scsi/scsi_device.h> | |
46 | #include <scsi/scsi_tcq.h> | |
47 | #include <scsi/scsi_transport.h> | |
48 | #include <linux/libata.h> | |
49 | #include <linux/hdreg.h> | |
50 | #include <linux/uaccess.h> | |
51 | #include <linux/suspend.h> | |
52 | #include <asm/unaligned.h> | |
53 | ||
54 | #include "libata.h" | |
55 | #include "libata-transport.h" | |
56 | ||
57 | #define ATA_SCSI_RBUF_SIZE 4096 | |
58 | ||
59 | static DEFINE_SPINLOCK(ata_scsi_rbuf_lock); | |
60 | static u8 ata_scsi_rbuf[ATA_SCSI_RBUF_SIZE]; | |
61 | ||
62 | typedef unsigned int (*ata_xlat_func_t)(struct ata_queued_cmd *qc); | |
63 | ||
64 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, | |
65 | const struct scsi_device *scsidev); | |
66 | static struct ata_device *ata_scsi_find_dev(struct ata_port *ap, | |
67 | const struct scsi_device *scsidev); | |
68 | ||
69 | #define RW_RECOVERY_MPAGE 0x1 | |
70 | #define RW_RECOVERY_MPAGE_LEN 12 | |
71 | #define CACHE_MPAGE 0x8 | |
72 | #define CACHE_MPAGE_LEN 20 | |
73 | #define CONTROL_MPAGE 0xa | |
74 | #define CONTROL_MPAGE_LEN 12 | |
75 | #define ALL_MPAGES 0x3f | |
76 | #define ALL_SUB_MPAGES 0xff | |
77 | ||
78 | ||
79 | static const u8 def_rw_recovery_mpage[RW_RECOVERY_MPAGE_LEN] = { | |
80 | RW_RECOVERY_MPAGE, | |
81 | RW_RECOVERY_MPAGE_LEN - 2, | |
82 | (1 << 7), /* AWRE */ | |
83 | 0, /* read retry count */ | |
84 | 0, 0, 0, 0, | |
85 | 0, /* write retry count */ | |
86 | 0, 0, 0 | |
87 | }; | |
88 | ||
89 | static const u8 def_cache_mpage[CACHE_MPAGE_LEN] = { | |
90 | CACHE_MPAGE, | |
91 | CACHE_MPAGE_LEN - 2, | |
92 | 0, /* contains WCE, needs to be 0 for logic */ | |
93 | 0, 0, 0, 0, 0, 0, 0, 0, 0, | |
94 | 0, /* contains DRA, needs to be 0 for logic */ | |
95 | 0, 0, 0, 0, 0, 0, 0 | |
96 | }; | |
97 | ||
98 | static const u8 def_control_mpage[CONTROL_MPAGE_LEN] = { | |
99 | CONTROL_MPAGE, | |
100 | CONTROL_MPAGE_LEN - 2, | |
101 | 2, /* DSENSE=0, GLTSD=1 */ | |
102 | 0, /* [QAM+QERR may be 1, see 05-359r1] */ | |
103 | 0, 0, 0, 0, 0xff, 0xff, | |
104 | 0, 30 /* extended self test time, see 05-359r1 */ | |
105 | }; | |
106 | ||
107 | static const char *ata_lpm_policy_names[] = { | |
108 | [ATA_LPM_UNKNOWN] = "max_performance", | |
109 | [ATA_LPM_MAX_POWER] = "max_performance", | |
110 | [ATA_LPM_MED_POWER] = "medium_power", | |
111 | [ATA_LPM_MIN_POWER] = "min_power", | |
112 | }; | |
113 | ||
114 | static ssize_t ata_scsi_lpm_store(struct device *device, | |
115 | struct device_attribute *attr, | |
116 | const char *buf, size_t count) | |
117 | { | |
118 | struct Scsi_Host *shost = class_to_shost(device); | |
119 | struct ata_port *ap = ata_shost_to_port(shost); | |
120 | struct ata_link *link; | |
121 | struct ata_device *dev; | |
122 | enum ata_lpm_policy policy; | |
123 | unsigned long flags; | |
124 | ||
125 | /* UNKNOWN is internal state, iterate from MAX_POWER */ | |
126 | for (policy = ATA_LPM_MAX_POWER; | |
127 | policy < ARRAY_SIZE(ata_lpm_policy_names); policy++) { | |
128 | const char *name = ata_lpm_policy_names[policy]; | |
129 | ||
130 | if (strncmp(name, buf, strlen(name)) == 0) | |
131 | break; | |
132 | } | |
133 | if (policy == ARRAY_SIZE(ata_lpm_policy_names)) | |
134 | return -EINVAL; | |
135 | ||
136 | spin_lock_irqsave(ap->lock, flags); | |
137 | ||
138 | ata_for_each_link(link, ap, EDGE) { | |
139 | ata_for_each_dev(dev, &ap->link, ENABLED) { | |
140 | if (dev->horkage & ATA_HORKAGE_NOLPM) { | |
141 | count = -EOPNOTSUPP; | |
142 | goto out_unlock; | |
143 | } | |
144 | } | |
145 | } | |
146 | ||
147 | ap->target_lpm_policy = policy; | |
148 | ata_port_schedule_eh(ap); | |
149 | out_unlock: | |
150 | spin_unlock_irqrestore(ap->lock, flags); | |
151 | return count; | |
152 | } | |
153 | ||
154 | static ssize_t ata_scsi_lpm_show(struct device *dev, | |
155 | struct device_attribute *attr, char *buf) | |
156 | { | |
157 | struct Scsi_Host *shost = class_to_shost(dev); | |
158 | struct ata_port *ap = ata_shost_to_port(shost); | |
159 | ||
160 | if (ap->target_lpm_policy >= ARRAY_SIZE(ata_lpm_policy_names)) | |
161 | return -EINVAL; | |
162 | ||
163 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
164 | ata_lpm_policy_names[ap->target_lpm_policy]); | |
165 | } | |
166 | DEVICE_ATTR(link_power_management_policy, S_IRUGO | S_IWUSR, | |
167 | ata_scsi_lpm_show, ata_scsi_lpm_store); | |
168 | EXPORT_SYMBOL_GPL(dev_attr_link_power_management_policy); | |
169 | ||
170 | static ssize_t ata_scsi_park_show(struct device *device, | |
171 | struct device_attribute *attr, char *buf) | |
172 | { | |
173 | struct scsi_device *sdev = to_scsi_device(device); | |
174 | struct ata_port *ap; | |
175 | struct ata_link *link; | |
176 | struct ata_device *dev; | |
177 | unsigned long flags, now; | |
178 | unsigned int uninitialized_var(msecs); | |
179 | int rc = 0; | |
180 | ||
181 | ap = ata_shost_to_port(sdev->host); | |
182 | ||
183 | spin_lock_irqsave(ap->lock, flags); | |
184 | dev = ata_scsi_find_dev(ap, sdev); | |
185 | if (!dev) { | |
186 | rc = -ENODEV; | |
187 | goto unlock; | |
188 | } | |
189 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { | |
190 | rc = -EOPNOTSUPP; | |
191 | goto unlock; | |
192 | } | |
193 | ||
194 | link = dev->link; | |
195 | now = jiffies; | |
196 | if (ap->pflags & ATA_PFLAG_EH_IN_PROGRESS && | |
197 | link->eh_context.unloaded_mask & (1 << dev->devno) && | |
198 | time_after(dev->unpark_deadline, now)) | |
199 | msecs = jiffies_to_msecs(dev->unpark_deadline - now); | |
200 | else | |
201 | msecs = 0; | |
202 | ||
203 | unlock: | |
204 | spin_unlock_irq(ap->lock); | |
205 | ||
206 | return rc ? rc : snprintf(buf, 20, "%u\n", msecs); | |
207 | } | |
208 | ||
209 | static ssize_t ata_scsi_park_store(struct device *device, | |
210 | struct device_attribute *attr, | |
211 | const char *buf, size_t len) | |
212 | { | |
213 | struct scsi_device *sdev = to_scsi_device(device); | |
214 | struct ata_port *ap; | |
215 | struct ata_device *dev; | |
216 | long int input; | |
217 | unsigned long flags; | |
218 | int rc; | |
219 | ||
220 | rc = kstrtol(buf, 10, &input); | |
221 | if (rc) | |
222 | return rc; | |
223 | if (input < -2) | |
224 | return -EINVAL; | |
225 | if (input > ATA_TMOUT_MAX_PARK) { | |
226 | rc = -EOVERFLOW; | |
227 | input = ATA_TMOUT_MAX_PARK; | |
228 | } | |
229 | ||
230 | ap = ata_shost_to_port(sdev->host); | |
231 | ||
232 | spin_lock_irqsave(ap->lock, flags); | |
233 | dev = ata_scsi_find_dev(ap, sdev); | |
234 | if (unlikely(!dev)) { | |
235 | rc = -ENODEV; | |
236 | goto unlock; | |
237 | } | |
238 | if (dev->class != ATA_DEV_ATA && | |
239 | dev->class != ATA_DEV_ZAC) { | |
240 | rc = -EOPNOTSUPP; | |
241 | goto unlock; | |
242 | } | |
243 | ||
244 | if (input >= 0) { | |
245 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { | |
246 | rc = -EOPNOTSUPP; | |
247 | goto unlock; | |
248 | } | |
249 | ||
250 | dev->unpark_deadline = ata_deadline(jiffies, input); | |
251 | dev->link->eh_info.dev_action[dev->devno] |= ATA_EH_PARK; | |
252 | ata_port_schedule_eh(ap); | |
253 | complete(&ap->park_req_pending); | |
254 | } else { | |
255 | switch (input) { | |
256 | case -1: | |
257 | dev->flags &= ~ATA_DFLAG_NO_UNLOAD; | |
258 | break; | |
259 | case -2: | |
260 | dev->flags |= ATA_DFLAG_NO_UNLOAD; | |
261 | break; | |
262 | } | |
263 | } | |
264 | unlock: | |
265 | spin_unlock_irqrestore(ap->lock, flags); | |
266 | ||
267 | return rc ? rc : len; | |
268 | } | |
269 | DEVICE_ATTR(unload_heads, S_IRUGO | S_IWUSR, | |
270 | ata_scsi_park_show, ata_scsi_park_store); | |
271 | EXPORT_SYMBOL_GPL(dev_attr_unload_heads); | |
272 | ||
273 | static void ata_scsi_set_sense(struct scsi_cmnd *cmd, u8 sk, u8 asc, u8 ascq) | |
274 | { | |
275 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; | |
276 | ||
277 | scsi_build_sense_buffer(0, cmd->sense_buffer, sk, asc, ascq); | |
278 | } | |
279 | ||
280 | static ssize_t | |
281 | ata_scsi_em_message_store(struct device *dev, struct device_attribute *attr, | |
282 | const char *buf, size_t count) | |
283 | { | |
284 | struct Scsi_Host *shost = class_to_shost(dev); | |
285 | struct ata_port *ap = ata_shost_to_port(shost); | |
286 | if (ap->ops->em_store && (ap->flags & ATA_FLAG_EM)) | |
287 | return ap->ops->em_store(ap, buf, count); | |
288 | return -EINVAL; | |
289 | } | |
290 | ||
291 | static ssize_t | |
292 | ata_scsi_em_message_show(struct device *dev, struct device_attribute *attr, | |
293 | char *buf) | |
294 | { | |
295 | struct Scsi_Host *shost = class_to_shost(dev); | |
296 | struct ata_port *ap = ata_shost_to_port(shost); | |
297 | ||
298 | if (ap->ops->em_show && (ap->flags & ATA_FLAG_EM)) | |
299 | return ap->ops->em_show(ap, buf); | |
300 | return -EINVAL; | |
301 | } | |
302 | DEVICE_ATTR(em_message, S_IRUGO | S_IWUSR, | |
303 | ata_scsi_em_message_show, ata_scsi_em_message_store); | |
304 | EXPORT_SYMBOL_GPL(dev_attr_em_message); | |
305 | ||
306 | static ssize_t | |
307 | ata_scsi_em_message_type_show(struct device *dev, struct device_attribute *attr, | |
308 | char *buf) | |
309 | { | |
310 | struct Scsi_Host *shost = class_to_shost(dev); | |
311 | struct ata_port *ap = ata_shost_to_port(shost); | |
312 | ||
313 | return snprintf(buf, 23, "%d\n", ap->em_message_type); | |
314 | } | |
315 | DEVICE_ATTR(em_message_type, S_IRUGO, | |
316 | ata_scsi_em_message_type_show, NULL); | |
317 | EXPORT_SYMBOL_GPL(dev_attr_em_message_type); | |
318 | ||
319 | static ssize_t | |
320 | ata_scsi_activity_show(struct device *dev, struct device_attribute *attr, | |
321 | char *buf) | |
322 | { | |
323 | struct scsi_device *sdev = to_scsi_device(dev); | |
324 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
325 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); | |
326 | ||
327 | if (atadev && ap->ops->sw_activity_show && | |
328 | (ap->flags & ATA_FLAG_SW_ACTIVITY)) | |
329 | return ap->ops->sw_activity_show(atadev, buf); | |
330 | return -EINVAL; | |
331 | } | |
332 | ||
333 | static ssize_t | |
334 | ata_scsi_activity_store(struct device *dev, struct device_attribute *attr, | |
335 | const char *buf, size_t count) | |
336 | { | |
337 | struct scsi_device *sdev = to_scsi_device(dev); | |
338 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
339 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); | |
340 | enum sw_activity val; | |
341 | int rc; | |
342 | ||
343 | if (atadev && ap->ops->sw_activity_store && | |
344 | (ap->flags & ATA_FLAG_SW_ACTIVITY)) { | |
345 | val = simple_strtoul(buf, NULL, 0); | |
346 | switch (val) { | |
347 | case OFF: case BLINK_ON: case BLINK_OFF: | |
348 | rc = ap->ops->sw_activity_store(atadev, val); | |
349 | if (!rc) | |
350 | return count; | |
351 | else | |
352 | return rc; | |
353 | } | |
354 | } | |
355 | return -EINVAL; | |
356 | } | |
357 | DEVICE_ATTR(sw_activity, S_IWUSR | S_IRUGO, ata_scsi_activity_show, | |
358 | ata_scsi_activity_store); | |
359 | EXPORT_SYMBOL_GPL(dev_attr_sw_activity); | |
360 | ||
361 | struct device_attribute *ata_common_sdev_attrs[] = { | |
362 | &dev_attr_unload_heads, | |
363 | NULL | |
364 | }; | |
365 | EXPORT_SYMBOL_GPL(ata_common_sdev_attrs); | |
366 | ||
367 | static void ata_scsi_invalid_field(struct scsi_cmnd *cmd) | |
368 | { | |
369 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
370 | /* "Invalid field in cbd" */ | |
371 | cmd->scsi_done(cmd); | |
372 | } | |
373 | ||
374 | /** | |
375 | * ata_std_bios_param - generic bios head/sector/cylinder calculator used by sd. | |
376 | * @sdev: SCSI device for which BIOS geometry is to be determined | |
377 | * @bdev: block device associated with @sdev | |
378 | * @capacity: capacity of SCSI device | |
379 | * @geom: location to which geometry will be output | |
380 | * | |
381 | * Generic bios head/sector/cylinder calculator | |
382 | * used by sd. Most BIOSes nowadays expect a XXX/255/16 (CHS) | |
383 | * mapping. Some situations may arise where the disk is not | |
384 | * bootable if this is not used. | |
385 | * | |
386 | * LOCKING: | |
387 | * Defined by the SCSI layer. We don't really care. | |
388 | * | |
389 | * RETURNS: | |
390 | * Zero. | |
391 | */ | |
392 | int ata_std_bios_param(struct scsi_device *sdev, struct block_device *bdev, | |
393 | sector_t capacity, int geom[]) | |
394 | { | |
395 | geom[0] = 255; | |
396 | geom[1] = 63; | |
397 | sector_div(capacity, 255*63); | |
398 | geom[2] = capacity; | |
399 | ||
400 | return 0; | |
401 | } | |
402 | ||
403 | /** | |
404 | * ata_scsi_unlock_native_capacity - unlock native capacity | |
405 | * @sdev: SCSI device to adjust device capacity for | |
406 | * | |
407 | * This function is called if a partition on @sdev extends beyond | |
408 | * the end of the device. It requests EH to unlock HPA. | |
409 | * | |
410 | * LOCKING: | |
411 | * Defined by the SCSI layer. Might sleep. | |
412 | */ | |
413 | void ata_scsi_unlock_native_capacity(struct scsi_device *sdev) | |
414 | { | |
415 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
416 | struct ata_device *dev; | |
417 | unsigned long flags; | |
418 | ||
419 | spin_lock_irqsave(ap->lock, flags); | |
420 | ||
421 | dev = ata_scsi_find_dev(ap, sdev); | |
422 | if (dev && dev->n_sectors < dev->n_native_sectors) { | |
423 | dev->flags |= ATA_DFLAG_UNLOCK_HPA; | |
424 | dev->link->eh_info.action |= ATA_EH_RESET; | |
425 | ata_port_schedule_eh(ap); | |
426 | } | |
427 | ||
428 | spin_unlock_irqrestore(ap->lock, flags); | |
429 | ata_port_wait_eh(ap); | |
430 | } | |
431 | ||
432 | /** | |
433 | * ata_get_identity - Handler for HDIO_GET_IDENTITY ioctl | |
434 | * @ap: target port | |
435 | * @sdev: SCSI device to get identify data for | |
436 | * @arg: User buffer area for identify data | |
437 | * | |
438 | * LOCKING: | |
439 | * Defined by the SCSI layer. We don't really care. | |
440 | * | |
441 | * RETURNS: | |
442 | * Zero on success, negative errno on error. | |
443 | */ | |
444 | static int ata_get_identity(struct ata_port *ap, struct scsi_device *sdev, | |
445 | void __user *arg) | |
446 | { | |
447 | struct ata_device *dev = ata_scsi_find_dev(ap, sdev); | |
448 | u16 __user *dst = arg; | |
449 | char buf[40]; | |
450 | ||
451 | if (!dev) | |
452 | return -ENOMSG; | |
453 | ||
454 | if (copy_to_user(dst, dev->id, ATA_ID_WORDS * sizeof(u16))) | |
455 | return -EFAULT; | |
456 | ||
457 | ata_id_string(dev->id, buf, ATA_ID_PROD, ATA_ID_PROD_LEN); | |
458 | if (copy_to_user(dst + ATA_ID_PROD, buf, ATA_ID_PROD_LEN)) | |
459 | return -EFAULT; | |
460 | ||
461 | ata_id_string(dev->id, buf, ATA_ID_FW_REV, ATA_ID_FW_REV_LEN); | |
462 | if (copy_to_user(dst + ATA_ID_FW_REV, buf, ATA_ID_FW_REV_LEN)) | |
463 | return -EFAULT; | |
464 | ||
465 | ata_id_string(dev->id, buf, ATA_ID_SERNO, ATA_ID_SERNO_LEN); | |
466 | if (copy_to_user(dst + ATA_ID_SERNO, buf, ATA_ID_SERNO_LEN)) | |
467 | return -EFAULT; | |
468 | ||
469 | return 0; | |
470 | } | |
471 | ||
472 | /** | |
473 | * ata_cmd_ioctl - Handler for HDIO_DRIVE_CMD ioctl | |
474 | * @scsidev: Device to which we are issuing command | |
475 | * @arg: User provided data for issuing command | |
476 | * | |
477 | * LOCKING: | |
478 | * Defined by the SCSI layer. We don't really care. | |
479 | * | |
480 | * RETURNS: | |
481 | * Zero on success, negative errno on error. | |
482 | */ | |
483 | int ata_cmd_ioctl(struct scsi_device *scsidev, void __user *arg) | |
484 | { | |
485 | int rc = 0; | |
486 | u8 scsi_cmd[MAX_COMMAND_SIZE]; | |
487 | u8 args[4], *argbuf = NULL, *sensebuf = NULL; | |
488 | int argsize = 0; | |
489 | enum dma_data_direction data_dir; | |
490 | int cmd_result; | |
491 | ||
492 | if (arg == NULL) | |
493 | return -EINVAL; | |
494 | ||
495 | if (copy_from_user(args, arg, sizeof(args))) | |
496 | return -EFAULT; | |
497 | ||
498 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); | |
499 | if (!sensebuf) | |
500 | return -ENOMEM; | |
501 | ||
502 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); | |
503 | ||
504 | if (args[3]) { | |
505 | argsize = ATA_SECT_SIZE * args[3]; | |
506 | argbuf = kmalloc(argsize, GFP_KERNEL); | |
507 | if (argbuf == NULL) { | |
508 | rc = -ENOMEM; | |
509 | goto error; | |
510 | } | |
511 | ||
512 | scsi_cmd[1] = (4 << 1); /* PIO Data-in */ | |
513 | scsi_cmd[2] = 0x0e; /* no off.line or cc, read from dev, | |
514 | block count in sector count field */ | |
515 | data_dir = DMA_FROM_DEVICE; | |
516 | } else { | |
517 | scsi_cmd[1] = (3 << 1); /* Non-data */ | |
518 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ | |
519 | data_dir = DMA_NONE; | |
520 | } | |
521 | ||
522 | scsi_cmd[0] = ATA_16; | |
523 | ||
524 | scsi_cmd[4] = args[2]; | |
525 | if (args[0] == ATA_CMD_SMART) { /* hack -- ide driver does this too */ | |
526 | scsi_cmd[6] = args[3]; | |
527 | scsi_cmd[8] = args[1]; | |
528 | scsi_cmd[10] = 0x4f; | |
529 | scsi_cmd[12] = 0xc2; | |
530 | } else { | |
531 | scsi_cmd[6] = args[1]; | |
532 | } | |
533 | scsi_cmd[14] = args[0]; | |
534 | ||
535 | /* Good values for timeout and retries? Values below | |
536 | from scsi_ioctl_send_command() for default case... */ | |
537 | cmd_result = scsi_execute(scsidev, scsi_cmd, data_dir, argbuf, argsize, | |
538 | sensebuf, (10*HZ), 5, 0, NULL); | |
539 | ||
540 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ | |
541 | u8 *desc = sensebuf + 8; | |
542 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ | |
543 | ||
544 | /* If we set cc then ATA pass-through will cause a | |
545 | * check condition even if no error. Filter that. */ | |
546 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { | |
547 | struct scsi_sense_hdr sshdr; | |
548 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, | |
549 | &sshdr); | |
550 | if (sshdr.sense_key == RECOVERED_ERROR && | |
551 | sshdr.asc == 0 && sshdr.ascq == 0x1d) | |
552 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; | |
553 | } | |
554 | ||
555 | /* Send userspace a few ATA registers (same as drivers/ide) */ | |
556 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ | |
557 | desc[0] == 0x09) { /* code is "ATA Descriptor" */ | |
558 | args[0] = desc[13]; /* status */ | |
559 | args[1] = desc[3]; /* error */ | |
560 | args[2] = desc[5]; /* sector count (0:7) */ | |
561 | if (copy_to_user(arg, args, sizeof(args))) | |
562 | rc = -EFAULT; | |
563 | } | |
564 | } | |
565 | ||
566 | ||
567 | if (cmd_result) { | |
568 | rc = -EIO; | |
569 | goto error; | |
570 | } | |
571 | ||
572 | if ((argbuf) | |
573 | && copy_to_user(arg + sizeof(args), argbuf, argsize)) | |
574 | rc = -EFAULT; | |
575 | error: | |
576 | kfree(sensebuf); | |
577 | kfree(argbuf); | |
578 | return rc; | |
579 | } | |
580 | ||
581 | /** | |
582 | * ata_task_ioctl - Handler for HDIO_DRIVE_TASK ioctl | |
583 | * @scsidev: Device to which we are issuing command | |
584 | * @arg: User provided data for issuing command | |
585 | * | |
586 | * LOCKING: | |
587 | * Defined by the SCSI layer. We don't really care. | |
588 | * | |
589 | * RETURNS: | |
590 | * Zero on success, negative errno on error. | |
591 | */ | |
592 | int ata_task_ioctl(struct scsi_device *scsidev, void __user *arg) | |
593 | { | |
594 | int rc = 0; | |
595 | u8 scsi_cmd[MAX_COMMAND_SIZE]; | |
596 | u8 args[7], *sensebuf = NULL; | |
597 | int cmd_result; | |
598 | ||
599 | if (arg == NULL) | |
600 | return -EINVAL; | |
601 | ||
602 | if (copy_from_user(args, arg, sizeof(args))) | |
603 | return -EFAULT; | |
604 | ||
605 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); | |
606 | if (!sensebuf) | |
607 | return -ENOMEM; | |
608 | ||
609 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); | |
610 | scsi_cmd[0] = ATA_16; | |
611 | scsi_cmd[1] = (3 << 1); /* Non-data */ | |
612 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ | |
613 | scsi_cmd[4] = args[1]; | |
614 | scsi_cmd[6] = args[2]; | |
615 | scsi_cmd[8] = args[3]; | |
616 | scsi_cmd[10] = args[4]; | |
617 | scsi_cmd[12] = args[5]; | |
618 | scsi_cmd[13] = args[6] & 0x4f; | |
619 | scsi_cmd[14] = args[0]; | |
620 | ||
621 | /* Good values for timeout and retries? Values below | |
622 | from scsi_ioctl_send_command() for default case... */ | |
623 | cmd_result = scsi_execute(scsidev, scsi_cmd, DMA_NONE, NULL, 0, | |
624 | sensebuf, (10*HZ), 5, 0, NULL); | |
625 | ||
626 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ | |
627 | u8 *desc = sensebuf + 8; | |
628 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ | |
629 | ||
630 | /* If we set cc then ATA pass-through will cause a | |
631 | * check condition even if no error. Filter that. */ | |
632 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { | |
633 | struct scsi_sense_hdr sshdr; | |
634 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, | |
635 | &sshdr); | |
636 | if (sshdr.sense_key == RECOVERED_ERROR && | |
637 | sshdr.asc == 0 && sshdr.ascq == 0x1d) | |
638 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; | |
639 | } | |
640 | ||
641 | /* Send userspace ATA registers */ | |
642 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ | |
643 | desc[0] == 0x09) {/* code is "ATA Descriptor" */ | |
644 | args[0] = desc[13]; /* status */ | |
645 | args[1] = desc[3]; /* error */ | |
646 | args[2] = desc[5]; /* sector count (0:7) */ | |
647 | args[3] = desc[7]; /* lbal */ | |
648 | args[4] = desc[9]; /* lbam */ | |
649 | args[5] = desc[11]; /* lbah */ | |
650 | args[6] = desc[12]; /* select */ | |
651 | if (copy_to_user(arg, args, sizeof(args))) | |
652 | rc = -EFAULT; | |
653 | } | |
654 | } | |
655 | ||
656 | if (cmd_result) { | |
657 | rc = -EIO; | |
658 | goto error; | |
659 | } | |
660 | ||
661 | error: | |
662 | kfree(sensebuf); | |
663 | return rc; | |
664 | } | |
665 | ||
666 | static int ata_ioc32(struct ata_port *ap) | |
667 | { | |
668 | if (ap->flags & ATA_FLAG_PIO_DMA) | |
669 | return 1; | |
670 | if (ap->pflags & ATA_PFLAG_PIO32) | |
671 | return 1; | |
672 | return 0; | |
673 | } | |
674 | ||
675 | int ata_sas_scsi_ioctl(struct ata_port *ap, struct scsi_device *scsidev, | |
676 | int cmd, void __user *arg) | |
677 | { | |
678 | int val = -EINVAL, rc = -EINVAL; | |
679 | unsigned long flags; | |
680 | ||
681 | switch (cmd) { | |
682 | case ATA_IOC_GET_IO32: | |
683 | spin_lock_irqsave(ap->lock, flags); | |
684 | val = ata_ioc32(ap); | |
685 | spin_unlock_irqrestore(ap->lock, flags); | |
686 | if (copy_to_user(arg, &val, 1)) | |
687 | return -EFAULT; | |
688 | return 0; | |
689 | ||
690 | case ATA_IOC_SET_IO32: | |
691 | val = (unsigned long) arg; | |
692 | rc = 0; | |
693 | spin_lock_irqsave(ap->lock, flags); | |
694 | if (ap->pflags & ATA_PFLAG_PIO32CHANGE) { | |
695 | if (val) | |
696 | ap->pflags |= ATA_PFLAG_PIO32; | |
697 | else | |
698 | ap->pflags &= ~ATA_PFLAG_PIO32; | |
699 | } else { | |
700 | if (val != ata_ioc32(ap)) | |
701 | rc = -EINVAL; | |
702 | } | |
703 | spin_unlock_irqrestore(ap->lock, flags); | |
704 | return rc; | |
705 | ||
706 | case HDIO_GET_IDENTITY: | |
707 | return ata_get_identity(ap, scsidev, arg); | |
708 | ||
709 | case HDIO_DRIVE_CMD: | |
710 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | |
711 | return -EACCES; | |
712 | return ata_cmd_ioctl(scsidev, arg); | |
713 | ||
714 | case HDIO_DRIVE_TASK: | |
715 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) | |
716 | return -EACCES; | |
717 | return ata_task_ioctl(scsidev, arg); | |
718 | ||
719 | default: | |
720 | rc = -ENOTTY; | |
721 | break; | |
722 | } | |
723 | ||
724 | return rc; | |
725 | } | |
726 | EXPORT_SYMBOL_GPL(ata_sas_scsi_ioctl); | |
727 | ||
728 | int ata_scsi_ioctl(struct scsi_device *scsidev, int cmd, void __user *arg) | |
729 | { | |
730 | return ata_sas_scsi_ioctl(ata_shost_to_port(scsidev->host), | |
731 | scsidev, cmd, arg); | |
732 | } | |
733 | EXPORT_SYMBOL_GPL(ata_scsi_ioctl); | |
734 | ||
735 | /** | |
736 | * ata_scsi_qc_new - acquire new ata_queued_cmd reference | |
737 | * @dev: ATA device to which the new command is attached | |
738 | * @cmd: SCSI command that originated this ATA command | |
739 | * | |
740 | * Obtain a reference to an unused ata_queued_cmd structure, | |
741 | * which is the basic libata structure representing a single | |
742 | * ATA command sent to the hardware. | |
743 | * | |
744 | * If a command was available, fill in the SCSI-specific | |
745 | * portions of the structure with information on the | |
746 | * current command. | |
747 | * | |
748 | * LOCKING: | |
749 | * spin_lock_irqsave(host lock) | |
750 | * | |
751 | * RETURNS: | |
752 | * Command allocated, or %NULL if none available. | |
753 | */ | |
754 | static struct ata_queued_cmd *ata_scsi_qc_new(struct ata_device *dev, | |
755 | struct scsi_cmnd *cmd) | |
756 | { | |
757 | struct ata_queued_cmd *qc; | |
758 | ||
759 | qc = ata_qc_new_init(dev); | |
760 | if (qc) { | |
761 | qc->scsicmd = cmd; | |
762 | qc->scsidone = cmd->scsi_done; | |
763 | ||
764 | qc->sg = scsi_sglist(cmd); | |
765 | qc->n_elem = scsi_sg_count(cmd); | |
766 | } else { | |
767 | cmd->result = (DID_OK << 16) | (QUEUE_FULL << 1); | |
768 | cmd->scsi_done(cmd); | |
769 | } | |
770 | ||
771 | return qc; | |
772 | } | |
773 | ||
774 | static void ata_qc_set_pc_nbytes(struct ata_queued_cmd *qc) | |
775 | { | |
776 | struct scsi_cmnd *scmd = qc->scsicmd; | |
777 | ||
778 | qc->extrabytes = scmd->request->extra_len; | |
779 | qc->nbytes = scsi_bufflen(scmd) + qc->extrabytes; | |
780 | } | |
781 | ||
782 | /** | |
783 | * ata_dump_status - user friendly display of error info | |
784 | * @id: id of the port in question | |
785 | * @tf: ptr to filled out taskfile | |
786 | * | |
787 | * Decode and dump the ATA error/status registers for the user so | |
788 | * that they have some idea what really happened at the non | |
789 | * make-believe layer. | |
790 | * | |
791 | * LOCKING: | |
792 | * inherited from caller | |
793 | */ | |
794 | static void ata_dump_status(unsigned id, struct ata_taskfile *tf) | |
795 | { | |
796 | u8 stat = tf->command, err = tf->feature; | |
797 | ||
798 | printk(KERN_WARNING "ata%u: status=0x%02x { ", id, stat); | |
799 | if (stat & ATA_BUSY) { | |
800 | printk("Busy }\n"); /* Data is not valid in this case */ | |
801 | } else { | |
802 | if (stat & 0x40) printk("DriveReady "); | |
803 | if (stat & 0x20) printk("DeviceFault "); | |
804 | if (stat & 0x10) printk("SeekComplete "); | |
805 | if (stat & 0x08) printk("DataRequest "); | |
806 | if (stat & 0x04) printk("CorrectedError "); | |
807 | if (stat & 0x02) printk("Index "); | |
808 | if (stat & 0x01) printk("Error "); | |
809 | printk("}\n"); | |
810 | ||
811 | if (err) { | |
812 | printk(KERN_WARNING "ata%u: error=0x%02x { ", id, err); | |
813 | if (err & 0x04) printk("DriveStatusError "); | |
814 | if (err & 0x80) { | |
815 | if (err & 0x04) printk("BadCRC "); | |
816 | else printk("Sector "); | |
817 | } | |
818 | if (err & 0x40) printk("UncorrectableError "); | |
819 | if (err & 0x10) printk("SectorIdNotFound "); | |
820 | if (err & 0x02) printk("TrackZeroNotFound "); | |
821 | if (err & 0x01) printk("AddrMarkNotFound "); | |
822 | printk("}\n"); | |
823 | } | |
824 | } | |
825 | } | |
826 | ||
827 | /** | |
828 | * ata_to_sense_error - convert ATA error to SCSI error | |
829 | * @id: ATA device number | |
830 | * @drv_stat: value contained in ATA status register | |
831 | * @drv_err: value contained in ATA error register | |
832 | * @sk: the sense key we'll fill out | |
833 | * @asc: the additional sense code we'll fill out | |
834 | * @ascq: the additional sense code qualifier we'll fill out | |
835 | * @verbose: be verbose | |
836 | * | |
837 | * Converts an ATA error into a SCSI error. Fill out pointers to | |
838 | * SK, ASC, and ASCQ bytes for later use in fixed or descriptor | |
839 | * format sense blocks. | |
840 | * | |
841 | * LOCKING: | |
842 | * spin_lock_irqsave(host lock) | |
843 | */ | |
844 | static void ata_to_sense_error(unsigned id, u8 drv_stat, u8 drv_err, u8 *sk, | |
845 | u8 *asc, u8 *ascq, int verbose) | |
846 | { | |
847 | int i; | |
848 | ||
849 | /* Based on the 3ware driver translation table */ | |
850 | static const unsigned char sense_table[][4] = { | |
851 | /* BBD|ECC|ID|MAR */ | |
852 | {0xd1, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command | |
853 | /* BBD|ECC|ID */ | |
854 | {0xd0, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command | |
855 | /* ECC|MC|MARK */ | |
856 | {0x61, HARDWARE_ERROR, 0x00, 0x00}, // Device fault Hardware error | |
857 | /* ICRC|ABRT */ /* NB: ICRC & !ABRT is BBD */ | |
858 | {0x84, ABORTED_COMMAND, 0x47, 0x00}, // Data CRC error SCSI parity error | |
859 | /* MC|ID|ABRT|TRK0|MARK */ | |
860 | {0x37, NOT_READY, 0x04, 0x00}, // Unit offline Not ready | |
861 | /* MCR|MARK */ | |
862 | {0x09, NOT_READY, 0x04, 0x00}, // Unrecovered disk error Not ready | |
863 | /* Bad address mark */ | |
864 | {0x01, MEDIUM_ERROR, 0x13, 0x00}, // Address mark not found Address mark not found for data field | |
865 | /* TRK0 */ | |
866 | {0x02, HARDWARE_ERROR, 0x00, 0x00}, // Track 0 not found Hardware error | |
867 | /* Abort: 0x04 is not translated here, see below */ | |
868 | /* Media change request */ | |
869 | {0x08, NOT_READY, 0x04, 0x00}, // Media change request FIXME: faking offline | |
870 | /* SRV/IDNF */ | |
871 | {0x10, ILLEGAL_REQUEST, 0x21, 0x00}, // ID not found Logical address out of range | |
872 | /* MC */ | |
873 | {0x20, UNIT_ATTENTION, 0x28, 0x00}, // Media Changed Not ready to ready change, medium may have changed | |
874 | /* ECC */ | |
875 | {0x40, MEDIUM_ERROR, 0x11, 0x04}, // Uncorrectable ECC error Unrecovered read error | |
876 | /* BBD - block marked bad */ | |
877 | {0x80, MEDIUM_ERROR, 0x11, 0x04}, // Block marked bad Medium error, unrecovered read error | |
878 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark | |
879 | }; | |
880 | static const unsigned char stat_table[][4] = { | |
881 | /* Must be first because BUSY means no other bits valid */ | |
882 | {0x80, ABORTED_COMMAND, 0x47, 0x00}, // Busy, fake parity for now | |
883 | {0x20, HARDWARE_ERROR, 0x44, 0x00}, // Device fault, internal target failure | |
884 | {0x08, ABORTED_COMMAND, 0x47, 0x00}, // Timed out in xfer, fake parity for now | |
885 | {0x04, RECOVERED_ERROR, 0x11, 0x00}, // Recovered ECC error Medium error, recovered | |
886 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark | |
887 | }; | |
888 | ||
889 | /* | |
890 | * Is this an error we can process/parse | |
891 | */ | |
892 | if (drv_stat & ATA_BUSY) { | |
893 | drv_err = 0; /* Ignore the err bits, they're invalid */ | |
894 | } | |
895 | ||
896 | if (drv_err) { | |
897 | /* Look for drv_err */ | |
898 | for (i = 0; sense_table[i][0] != 0xFF; i++) { | |
899 | /* Look for best matches first */ | |
900 | if ((sense_table[i][0] & drv_err) == | |
901 | sense_table[i][0]) { | |
902 | *sk = sense_table[i][1]; | |
903 | *asc = sense_table[i][2]; | |
904 | *ascq = sense_table[i][3]; | |
905 | goto translate_done; | |
906 | } | |
907 | } | |
908 | } | |
909 | ||
910 | /* | |
911 | * Fall back to interpreting status bits. Note that if the drv_err | |
912 | * has only the ABRT bit set, we decode drv_stat. ABRT by itself | |
913 | * is not descriptive enough. | |
914 | */ | |
915 | for (i = 0; stat_table[i][0] != 0xFF; i++) { | |
916 | if (stat_table[i][0] & drv_stat) { | |
917 | *sk = stat_table[i][1]; | |
918 | *asc = stat_table[i][2]; | |
919 | *ascq = stat_table[i][3]; | |
920 | goto translate_done; | |
921 | } | |
922 | } | |
923 | ||
924 | /* | |
925 | * We need a sensible error return here, which is tricky, and one | |
926 | * that won't cause people to do things like return a disk wrongly. | |
927 | */ | |
928 | *sk = ABORTED_COMMAND; | |
929 | *asc = 0x00; | |
930 | *ascq = 0x00; | |
931 | ||
932 | translate_done: | |
933 | if (verbose) | |
934 | printk(KERN_ERR "ata%u: translated ATA stat/err 0x%02x/%02x " | |
935 | "to SCSI SK/ASC/ASCQ 0x%x/%02x/%02x\n", | |
936 | id, drv_stat, drv_err, *sk, *asc, *ascq); | |
937 | return; | |
938 | } | |
939 | ||
940 | /* | |
941 | * ata_gen_passthru_sense - Generate check condition sense block. | |
942 | * @qc: Command that completed. | |
943 | * | |
944 | * This function is specific to the ATA descriptor format sense | |
945 | * block specified for the ATA pass through commands. Regardless | |
946 | * of whether the command errored or not, return a sense | |
947 | * block. Copy all controller registers into the sense | |
948 | * block. If there was no error, we get the request from an ATA | |
949 | * passthrough command, so we use the following sense data: | |
950 | * sk = RECOVERED ERROR | |
951 | * asc,ascq = ATA PASS-THROUGH INFORMATION AVAILABLE | |
952 | * | |
953 | * | |
954 | * LOCKING: | |
955 | * None. | |
956 | */ | |
957 | static void ata_gen_passthru_sense(struct ata_queued_cmd *qc) | |
958 | { | |
959 | struct scsi_cmnd *cmd = qc->scsicmd; | |
960 | struct ata_taskfile *tf = &qc->result_tf; | |
961 | unsigned char *sb = cmd->sense_buffer; | |
962 | unsigned char *desc = sb + 8; | |
963 | int verbose = qc->ap->ops->error_handler == NULL; | |
964 | ||
965 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); | |
966 | ||
967 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; | |
968 | ||
969 | /* | |
970 | * Use ata_to_sense_error() to map status register bits | |
971 | * onto sense key, asc & ascq. | |
972 | */ | |
973 | if (qc->err_mask || | |
974 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { | |
975 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, | |
976 | &sb[1], &sb[2], &sb[3], verbose); | |
977 | sb[1] &= 0x0f; | |
978 | } else { | |
979 | sb[1] = RECOVERED_ERROR; | |
980 | sb[2] = 0; | |
981 | sb[3] = 0x1D; | |
982 | } | |
983 | ||
984 | /* | |
985 | * Sense data is current and format is descriptor. | |
986 | */ | |
987 | sb[0] = 0x72; | |
988 | ||
989 | desc[0] = 0x09; | |
990 | ||
991 | /* set length of additional sense data */ | |
992 | sb[7] = 14; | |
993 | desc[1] = 12; | |
994 | ||
995 | /* | |
996 | * Copy registers into sense buffer. | |
997 | */ | |
998 | desc[2] = 0x00; | |
999 | desc[3] = tf->feature; /* == error reg */ | |
1000 | desc[5] = tf->nsect; | |
1001 | desc[7] = tf->lbal; | |
1002 | desc[9] = tf->lbam; | |
1003 | desc[11] = tf->lbah; | |
1004 | desc[12] = tf->device; | |
1005 | desc[13] = tf->command; /* == status reg */ | |
1006 | ||
1007 | /* | |
1008 | * Fill in Extend bit, and the high order bytes | |
1009 | * if applicable. | |
1010 | */ | |
1011 | if (tf->flags & ATA_TFLAG_LBA48) { | |
1012 | desc[2] |= 0x01; | |
1013 | desc[4] = tf->hob_nsect; | |
1014 | desc[6] = tf->hob_lbal; | |
1015 | desc[8] = tf->hob_lbam; | |
1016 | desc[10] = tf->hob_lbah; | |
1017 | } | |
1018 | } | |
1019 | ||
1020 | /** | |
1021 | * ata_gen_ata_sense - generate a SCSI fixed sense block | |
1022 | * @qc: Command that we are erroring out | |
1023 | * | |
1024 | * Generate sense block for a failed ATA command @qc. Descriptor | |
1025 | * format is used to accommodate LBA48 block address. | |
1026 | * | |
1027 | * LOCKING: | |
1028 | * None. | |
1029 | */ | |
1030 | static void ata_gen_ata_sense(struct ata_queued_cmd *qc) | |
1031 | { | |
1032 | struct ata_device *dev = qc->dev; | |
1033 | struct scsi_cmnd *cmd = qc->scsicmd; | |
1034 | struct ata_taskfile *tf = &qc->result_tf; | |
1035 | unsigned char *sb = cmd->sense_buffer; | |
1036 | unsigned char *desc = sb + 8; | |
1037 | int verbose = qc->ap->ops->error_handler == NULL; | |
1038 | u64 block; | |
1039 | ||
1040 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); | |
1041 | ||
1042 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; | |
1043 | ||
1044 | /* sense data is current and format is descriptor */ | |
1045 | sb[0] = 0x72; | |
1046 | ||
1047 | /* Use ata_to_sense_error() to map status register bits | |
1048 | * onto sense key, asc & ascq. | |
1049 | */ | |
1050 | if (qc->err_mask || | |
1051 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { | |
1052 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, | |
1053 | &sb[1], &sb[2], &sb[3], verbose); | |
1054 | sb[1] &= 0x0f; | |
1055 | } | |
1056 | ||
1057 | block = ata_tf_read_block(&qc->result_tf, dev); | |
1058 | ||
1059 | /* information sense data descriptor */ | |
1060 | sb[7] = 12; | |
1061 | desc[0] = 0x00; | |
1062 | desc[1] = 10; | |
1063 | ||
1064 | desc[2] |= 0x80; /* valid */ | |
1065 | desc[6] = block >> 40; | |
1066 | desc[7] = block >> 32; | |
1067 | desc[8] = block >> 24; | |
1068 | desc[9] = block >> 16; | |
1069 | desc[10] = block >> 8; | |
1070 | desc[11] = block; | |
1071 | } | |
1072 | ||
1073 | static void ata_scsi_sdev_config(struct scsi_device *sdev) | |
1074 | { | |
1075 | sdev->use_10_for_rw = 1; | |
1076 | sdev->use_10_for_ms = 1; | |
1077 | sdev->no_report_opcodes = 1; | |
1078 | sdev->no_write_same = 1; | |
1079 | ||
1080 | /* Schedule policy is determined by ->qc_defer() callback and | |
1081 | * it needs to see every deferred qc. Set dev_blocked to 1 to | |
1082 | * prevent SCSI midlayer from automatically deferring | |
1083 | * requests. | |
1084 | */ | |
1085 | sdev->max_device_blocked = 1; | |
1086 | } | |
1087 | ||
1088 | /** | |
1089 | * atapi_drain_needed - Check whether data transfer may overflow | |
1090 | * @rq: request to be checked | |
1091 | * | |
1092 | * ATAPI commands which transfer variable length data to host | |
1093 | * might overflow due to application error or hardare bug. This | |
1094 | * function checks whether overflow should be drained and ignored | |
1095 | * for @request. | |
1096 | * | |
1097 | * LOCKING: | |
1098 | * None. | |
1099 | * | |
1100 | * RETURNS: | |
1101 | * 1 if ; otherwise, 0. | |
1102 | */ | |
1103 | static int atapi_drain_needed(struct request *rq) | |
1104 | { | |
1105 | if (likely(rq->cmd_type != REQ_TYPE_BLOCK_PC)) | |
1106 | return 0; | |
1107 | ||
1108 | if (!blk_rq_bytes(rq) || (rq->cmd_flags & REQ_WRITE)) | |
1109 | return 0; | |
1110 | ||
1111 | return atapi_cmd_type(rq->cmd[0]) == ATAPI_MISC; | |
1112 | } | |
1113 | ||
1114 | static int ata_scsi_dev_config(struct scsi_device *sdev, | |
1115 | struct ata_device *dev) | |
1116 | { | |
1117 | struct request_queue *q = sdev->request_queue; | |
1118 | ||
1119 | if (!ata_id_has_unload(dev->id)) | |
1120 | dev->flags |= ATA_DFLAG_NO_UNLOAD; | |
1121 | ||
1122 | /* configure max sectors */ | |
1123 | blk_queue_max_hw_sectors(q, dev->max_sectors); | |
1124 | ||
1125 | if (dev->class == ATA_DEV_ATAPI) { | |
1126 | void *buf; | |
1127 | ||
1128 | sdev->sector_size = ATA_SECT_SIZE; | |
1129 | ||
1130 | /* set DMA padding */ | |
1131 | blk_queue_update_dma_pad(q, ATA_DMA_PAD_SZ - 1); | |
1132 | ||
1133 | /* configure draining */ | |
1134 | buf = kmalloc(ATAPI_MAX_DRAIN, q->bounce_gfp | GFP_KERNEL); | |
1135 | if (!buf) { | |
1136 | ata_dev_err(dev, "drain buffer allocation failed\n"); | |
1137 | return -ENOMEM; | |
1138 | } | |
1139 | ||
1140 | blk_queue_dma_drain(q, atapi_drain_needed, buf, ATAPI_MAX_DRAIN); | |
1141 | } else { | |
1142 | sdev->sector_size = ata_id_logical_sector_size(dev->id); | |
1143 | sdev->manage_start_stop = 1; | |
1144 | } | |
1145 | ||
1146 | /* | |
1147 | * ata_pio_sectors() expects buffer for each sector to not cross | |
1148 | * page boundary. Enforce it by requiring buffers to be sector | |
1149 | * aligned, which works iff sector_size is not larger than | |
1150 | * PAGE_SIZE. ATAPI devices also need the alignment as | |
1151 | * IDENTIFY_PACKET is executed as ATA_PROT_PIO. | |
1152 | */ | |
1153 | if (sdev->sector_size > PAGE_SIZE) | |
1154 | ata_dev_warn(dev, | |
1155 | "sector_size=%u > PAGE_SIZE, PIO may malfunction\n", | |
1156 | sdev->sector_size); | |
1157 | ||
1158 | blk_queue_update_dma_alignment(q, sdev->sector_size - 1); | |
1159 | ||
1160 | if (dev->flags & ATA_DFLAG_AN) | |
1161 | set_bit(SDEV_EVT_MEDIA_CHANGE, sdev->supported_events); | |
1162 | ||
1163 | if (dev->flags & ATA_DFLAG_NCQ) { | |
1164 | int depth; | |
1165 | ||
1166 | depth = min(sdev->host->can_queue, ata_id_queue_depth(dev->id)); | |
1167 | depth = min(ATA_MAX_QUEUE - 1, depth); | |
1168 | scsi_change_queue_depth(sdev, depth); | |
1169 | } | |
1170 | ||
1171 | blk_queue_flush_queueable(q, false); | |
1172 | ||
1173 | dev->sdev = sdev; | |
1174 | return 0; | |
1175 | } | |
1176 | ||
1177 | /** | |
1178 | * ata_scsi_slave_config - Set SCSI device attributes | |
1179 | * @sdev: SCSI device to examine | |
1180 | * | |
1181 | * This is called before we actually start reading | |
1182 | * and writing to the device, to configure certain | |
1183 | * SCSI mid-layer behaviors. | |
1184 | * | |
1185 | * LOCKING: | |
1186 | * Defined by SCSI layer. We don't really care. | |
1187 | */ | |
1188 | ||
1189 | int ata_scsi_slave_config(struct scsi_device *sdev) | |
1190 | { | |
1191 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
1192 | struct ata_device *dev = __ata_scsi_find_dev(ap, sdev); | |
1193 | int rc = 0; | |
1194 | ||
1195 | ata_scsi_sdev_config(sdev); | |
1196 | ||
1197 | if (dev) | |
1198 | rc = ata_scsi_dev_config(sdev, dev); | |
1199 | ||
1200 | return rc; | |
1201 | } | |
1202 | ||
1203 | /** | |
1204 | * ata_scsi_slave_destroy - SCSI device is about to be destroyed | |
1205 | * @sdev: SCSI device to be destroyed | |
1206 | * | |
1207 | * @sdev is about to be destroyed for hot/warm unplugging. If | |
1208 | * this unplugging was initiated by libata as indicated by NULL | |
1209 | * dev->sdev, this function doesn't have to do anything. | |
1210 | * Otherwise, SCSI layer initiated warm-unplug is in progress. | |
1211 | * Clear dev->sdev, schedule the device for ATA detach and invoke | |
1212 | * EH. | |
1213 | * | |
1214 | * LOCKING: | |
1215 | * Defined by SCSI layer. We don't really care. | |
1216 | */ | |
1217 | void ata_scsi_slave_destroy(struct scsi_device *sdev) | |
1218 | { | |
1219 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
1220 | struct request_queue *q = sdev->request_queue; | |
1221 | unsigned long flags; | |
1222 | struct ata_device *dev; | |
1223 | ||
1224 | if (!ap->ops->error_handler) | |
1225 | return; | |
1226 | ||
1227 | spin_lock_irqsave(ap->lock, flags); | |
1228 | dev = __ata_scsi_find_dev(ap, sdev); | |
1229 | if (dev && dev->sdev) { | |
1230 | /* SCSI device already in CANCEL state, no need to offline it */ | |
1231 | dev->sdev = NULL; | |
1232 | dev->flags |= ATA_DFLAG_DETACH; | |
1233 | ata_port_schedule_eh(ap); | |
1234 | } | |
1235 | spin_unlock_irqrestore(ap->lock, flags); | |
1236 | ||
1237 | kfree(q->dma_drain_buffer); | |
1238 | q->dma_drain_buffer = NULL; | |
1239 | q->dma_drain_size = 0; | |
1240 | } | |
1241 | ||
1242 | /** | |
1243 | * __ata_change_queue_depth - helper for ata_scsi_change_queue_depth | |
1244 | * @ap: ATA port to which the device change the queue depth | |
1245 | * @sdev: SCSI device to configure queue depth for | |
1246 | * @queue_depth: new queue depth | |
1247 | * | |
1248 | * libsas and libata have different approaches for associating a sdev to | |
1249 | * its ata_port. | |
1250 | * | |
1251 | */ | |
1252 | int __ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, | |
1253 | int queue_depth) | |
1254 | { | |
1255 | struct ata_device *dev; | |
1256 | unsigned long flags; | |
1257 | ||
1258 | if (queue_depth < 1 || queue_depth == sdev->queue_depth) | |
1259 | return sdev->queue_depth; | |
1260 | ||
1261 | dev = ata_scsi_find_dev(ap, sdev); | |
1262 | if (!dev || !ata_dev_enabled(dev)) | |
1263 | return sdev->queue_depth; | |
1264 | ||
1265 | /* NCQ enabled? */ | |
1266 | spin_lock_irqsave(ap->lock, flags); | |
1267 | dev->flags &= ~ATA_DFLAG_NCQ_OFF; | |
1268 | if (queue_depth == 1 || !ata_ncq_enabled(dev)) { | |
1269 | dev->flags |= ATA_DFLAG_NCQ_OFF; | |
1270 | queue_depth = 1; | |
1271 | } | |
1272 | spin_unlock_irqrestore(ap->lock, flags); | |
1273 | ||
1274 | /* limit and apply queue depth */ | |
1275 | queue_depth = min(queue_depth, sdev->host->can_queue); | |
1276 | queue_depth = min(queue_depth, ata_id_queue_depth(dev->id)); | |
1277 | queue_depth = min(queue_depth, ATA_MAX_QUEUE - 1); | |
1278 | ||
1279 | if (sdev->queue_depth == queue_depth) | |
1280 | return -EINVAL; | |
1281 | ||
1282 | return scsi_change_queue_depth(sdev, queue_depth); | |
1283 | } | |
1284 | ||
1285 | /** | |
1286 | * ata_scsi_change_queue_depth - SCSI callback for queue depth config | |
1287 | * @sdev: SCSI device to configure queue depth for | |
1288 | * @queue_depth: new queue depth | |
1289 | * | |
1290 | * This is libata standard hostt->change_queue_depth callback. | |
1291 | * SCSI will call into this callback when user tries to set queue | |
1292 | * depth via sysfs. | |
1293 | * | |
1294 | * LOCKING: | |
1295 | * SCSI layer (we don't care) | |
1296 | * | |
1297 | * RETURNS: | |
1298 | * Newly configured queue depth. | |
1299 | */ | |
1300 | int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) | |
1301 | { | |
1302 | struct ata_port *ap = ata_shost_to_port(sdev->host); | |
1303 | ||
1304 | return __ata_change_queue_depth(ap, sdev, queue_depth); | |
1305 | } | |
1306 | ||
1307 | /** | |
1308 | * ata_scsi_start_stop_xlat - Translate SCSI START STOP UNIT command | |
1309 | * @qc: Storage for translated ATA taskfile | |
1310 | * | |
1311 | * Sets up an ATA taskfile to issue STANDBY (to stop) or READ VERIFY | |
1312 | * (to start). Perhaps these commands should be preceded by | |
1313 | * CHECK POWER MODE to see what power mode the device is already in. | |
1314 | * [See SAT revision 5 at www.t10.org] | |
1315 | * | |
1316 | * LOCKING: | |
1317 | * spin_lock_irqsave(host lock) | |
1318 | * | |
1319 | * RETURNS: | |
1320 | * Zero on success, non-zero on error. | |
1321 | */ | |
1322 | static unsigned int ata_scsi_start_stop_xlat(struct ata_queued_cmd *qc) | |
1323 | { | |
1324 | struct scsi_cmnd *scmd = qc->scsicmd; | |
1325 | struct ata_taskfile *tf = &qc->tf; | |
1326 | const u8 *cdb = scmd->cmnd; | |
1327 | ||
1328 | if (scmd->cmd_len < 5) | |
1329 | goto invalid_fld; | |
1330 | ||
1331 | tf->flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; | |
1332 | tf->protocol = ATA_PROT_NODATA; | |
1333 | if (cdb[1] & 0x1) { | |
1334 | ; /* ignore IMMED bit, violates sat-r05 */ | |
1335 | } | |
1336 | if (cdb[4] & 0x2) | |
1337 | goto invalid_fld; /* LOEJ bit set not supported */ | |
1338 | if (((cdb[4] >> 4) & 0xf) != 0) | |
1339 | goto invalid_fld; /* power conditions not supported */ | |
1340 | ||
1341 | if (cdb[4] & 0x1) { | |
1342 | tf->nsect = 1; /* 1 sector, lba=0 */ | |
1343 | ||
1344 | if (qc->dev->flags & ATA_DFLAG_LBA) { | |
1345 | tf->flags |= ATA_TFLAG_LBA; | |
1346 | ||
1347 | tf->lbah = 0x0; | |
1348 | tf->lbam = 0x0; | |
1349 | tf->lbal = 0x0; | |
1350 | tf->device |= ATA_LBA; | |
1351 | } else { | |
1352 | /* CHS */ | |
1353 | tf->lbal = 0x1; /* sect */ | |
1354 | tf->lbam = 0x0; /* cyl low */ | |
1355 | tf->lbah = 0x0; /* cyl high */ | |
1356 | } | |
1357 | ||
1358 | tf->command = ATA_CMD_VERIFY; /* READ VERIFY */ | |
1359 | } else { | |
1360 | /* Some odd clown BIOSen issue spindown on power off (ACPI S4 | |
1361 | * or S5) causing some drives to spin up and down again. | |
1362 | */ | |
1363 | if ((qc->ap->flags & ATA_FLAG_NO_POWEROFF_SPINDOWN) && | |
1364 | system_state == SYSTEM_POWER_OFF) | |
1365 | goto skip; | |
1366 | ||
1367 | if ((qc->ap->flags & ATA_FLAG_NO_HIBERNATE_SPINDOWN) && | |
1368 | system_entering_hibernation()) | |
1369 | goto skip; | |
1370 | ||
1371 | /* Issue ATA STANDBY IMMEDIATE command */ | |
1372 | tf->command = ATA_CMD_STANDBYNOW1; | |
1373 | } | |
1374 | ||
1375 | /* | |
1376 | * Standby and Idle condition timers could be implemented but that | |
1377 | * would require libata to implement the Power condition mode page | |
1378 | * and allow the user to change it. Changing mode pages requires | |
1379 | * MODE SELECT to be implemented. | |
1380 | */ | |
1381 | ||
1382 | return 0; | |
1383 | ||
1384 | invalid_fld: | |
1385 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
1386 | /* "Invalid field in cbd" */ | |
1387 | return 1; | |
1388 | skip: | |
1389 | scmd->result = SAM_STAT_GOOD; | |
1390 | return 1; | |
1391 | } | |
1392 | ||
1393 | ||
1394 | /** | |
1395 | * ata_scsi_flush_xlat - Translate SCSI SYNCHRONIZE CACHE command | |
1396 | * @qc: Storage for translated ATA taskfile | |
1397 | * | |
1398 | * Sets up an ATA taskfile to issue FLUSH CACHE or | |
1399 | * FLUSH CACHE EXT. | |
1400 | * | |
1401 | * LOCKING: | |
1402 | * spin_lock_irqsave(host lock) | |
1403 | * | |
1404 | * RETURNS: | |
1405 | * Zero on success, non-zero on error. | |
1406 | */ | |
1407 | static unsigned int ata_scsi_flush_xlat(struct ata_queued_cmd *qc) | |
1408 | { | |
1409 | struct ata_taskfile *tf = &qc->tf; | |
1410 | ||
1411 | tf->flags |= ATA_TFLAG_DEVICE; | |
1412 | tf->protocol = ATA_PROT_NODATA; | |
1413 | ||
1414 | if (qc->dev->flags & ATA_DFLAG_FLUSH_EXT) | |
1415 | tf->command = ATA_CMD_FLUSH_EXT; | |
1416 | else | |
1417 | tf->command = ATA_CMD_FLUSH; | |
1418 | ||
1419 | /* flush is critical for IO integrity, consider it an IO command */ | |
1420 | qc->flags |= ATA_QCFLAG_IO; | |
1421 | ||
1422 | return 0; | |
1423 | } | |
1424 | ||
1425 | /** | |
1426 | * scsi_6_lba_len - Get LBA and transfer length | |
1427 | * @cdb: SCSI command to translate | |
1428 | * | |
1429 | * Calculate LBA and transfer length for 6-byte commands. | |
1430 | * | |
1431 | * RETURNS: | |
1432 | * @plba: the LBA | |
1433 | * @plen: the transfer length | |
1434 | */ | |
1435 | static void scsi_6_lba_len(const u8 *cdb, u64 *plba, u32 *plen) | |
1436 | { | |
1437 | u64 lba = 0; | |
1438 | u32 len; | |
1439 | ||
1440 | VPRINTK("six-byte command\n"); | |
1441 | ||
1442 | lba |= ((u64)(cdb[1] & 0x1f)) << 16; | |
1443 | lba |= ((u64)cdb[2]) << 8; | |
1444 | lba |= ((u64)cdb[3]); | |
1445 | ||
1446 | len = cdb[4]; | |
1447 | ||
1448 | *plba = lba; | |
1449 | *plen = len; | |
1450 | } | |
1451 | ||
1452 | /** | |
1453 | * scsi_10_lba_len - Get LBA and transfer length | |
1454 | * @cdb: SCSI command to translate | |
1455 | * | |
1456 | * Calculate LBA and transfer length for 10-byte commands. | |
1457 | * | |
1458 | * RETURNS: | |
1459 | * @plba: the LBA | |
1460 | * @plen: the transfer length | |
1461 | */ | |
1462 | static void scsi_10_lba_len(const u8 *cdb, u64 *plba, u32 *plen) | |
1463 | { | |
1464 | u64 lba = 0; | |
1465 | u32 len = 0; | |
1466 | ||
1467 | VPRINTK("ten-byte command\n"); | |
1468 | ||
1469 | lba |= ((u64)cdb[2]) << 24; | |
1470 | lba |= ((u64)cdb[3]) << 16; | |
1471 | lba |= ((u64)cdb[4]) << 8; | |
1472 | lba |= ((u64)cdb[5]); | |
1473 | ||
1474 | len |= ((u32)cdb[7]) << 8; | |
1475 | len |= ((u32)cdb[8]); | |
1476 | ||
1477 | *plba = lba; | |
1478 | *plen = len; | |
1479 | } | |
1480 | ||
1481 | /** | |
1482 | * scsi_16_lba_len - Get LBA and transfer length | |
1483 | * @cdb: SCSI command to translate | |
1484 | * | |
1485 | * Calculate LBA and transfer length for 16-byte commands. | |
1486 | * | |
1487 | * RETURNS: | |
1488 | * @plba: the LBA | |
1489 | * @plen: the transfer length | |
1490 | */ | |
1491 | static void scsi_16_lba_len(const u8 *cdb, u64 *plba, u32 *plen) | |
1492 | { | |
1493 | u64 lba = 0; | |
1494 | u32 len = 0; | |
1495 | ||
1496 | VPRINTK("sixteen-byte command\n"); | |
1497 | ||
1498 | lba |= ((u64)cdb[2]) << 56; | |
1499 | lba |= ((u64)cdb[3]) << 48; | |
1500 | lba |= ((u64)cdb[4]) << 40; | |
1501 | lba |= ((u64)cdb[5]) << 32; | |
1502 | lba |= ((u64)cdb[6]) << 24; | |
1503 | lba |= ((u64)cdb[7]) << 16; | |
1504 | lba |= ((u64)cdb[8]) << 8; | |
1505 | lba |= ((u64)cdb[9]); | |
1506 | ||
1507 | len |= ((u32)cdb[10]) << 24; | |
1508 | len |= ((u32)cdb[11]) << 16; | |
1509 | len |= ((u32)cdb[12]) << 8; | |
1510 | len |= ((u32)cdb[13]); | |
1511 | ||
1512 | *plba = lba; | |
1513 | *plen = len; | |
1514 | } | |
1515 | ||
1516 | /** | |
1517 | * ata_scsi_verify_xlat - Translate SCSI VERIFY command into an ATA one | |
1518 | * @qc: Storage for translated ATA taskfile | |
1519 | * | |
1520 | * Converts SCSI VERIFY command to an ATA READ VERIFY command. | |
1521 | * | |
1522 | * LOCKING: | |
1523 | * spin_lock_irqsave(host lock) | |
1524 | * | |
1525 | * RETURNS: | |
1526 | * Zero on success, non-zero on error. | |
1527 | */ | |
1528 | static unsigned int ata_scsi_verify_xlat(struct ata_queued_cmd *qc) | |
1529 | { | |
1530 | struct scsi_cmnd *scmd = qc->scsicmd; | |
1531 | struct ata_taskfile *tf = &qc->tf; | |
1532 | struct ata_device *dev = qc->dev; | |
1533 | u64 dev_sectors = qc->dev->n_sectors; | |
1534 | const u8 *cdb = scmd->cmnd; | |
1535 | u64 block; | |
1536 | u32 n_block; | |
1537 | ||
1538 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; | |
1539 | tf->protocol = ATA_PROT_NODATA; | |
1540 | ||
1541 | if (cdb[0] == VERIFY) { | |
1542 | if (scmd->cmd_len < 10) | |
1543 | goto invalid_fld; | |
1544 | scsi_10_lba_len(cdb, &block, &n_block); | |
1545 | } else if (cdb[0] == VERIFY_16) { | |
1546 | if (scmd->cmd_len < 16) | |
1547 | goto invalid_fld; | |
1548 | scsi_16_lba_len(cdb, &block, &n_block); | |
1549 | } else | |
1550 | goto invalid_fld; | |
1551 | ||
1552 | if (!n_block) | |
1553 | goto nothing_to_do; | |
1554 | if (block >= dev_sectors) | |
1555 | goto out_of_range; | |
1556 | if ((block + n_block) > dev_sectors) | |
1557 | goto out_of_range; | |
1558 | ||
1559 | if (dev->flags & ATA_DFLAG_LBA) { | |
1560 | tf->flags |= ATA_TFLAG_LBA; | |
1561 | ||
1562 | if (lba_28_ok(block, n_block)) { | |
1563 | /* use LBA28 */ | |
1564 | tf->command = ATA_CMD_VERIFY; | |
1565 | tf->device |= (block >> 24) & 0xf; | |
1566 | } else if (lba_48_ok(block, n_block)) { | |
1567 | if (!(dev->flags & ATA_DFLAG_LBA48)) | |
1568 | goto out_of_range; | |
1569 | ||
1570 | /* use LBA48 */ | |
1571 | tf->flags |= ATA_TFLAG_LBA48; | |
1572 | tf->command = ATA_CMD_VERIFY_EXT; | |
1573 | ||
1574 | tf->hob_nsect = (n_block >> 8) & 0xff; | |
1575 | ||
1576 | tf->hob_lbah = (block >> 40) & 0xff; | |
1577 | tf->hob_lbam = (block >> 32) & 0xff; | |
1578 | tf->hob_lbal = (block >> 24) & 0xff; | |
1579 | } else | |
1580 | /* request too large even for LBA48 */ | |
1581 | goto out_of_range; | |
1582 | ||
1583 | tf->nsect = n_block & 0xff; | |
1584 | ||
1585 | tf->lbah = (block >> 16) & 0xff; | |
1586 | tf->lbam = (block >> 8) & 0xff; | |
1587 | tf->lbal = block & 0xff; | |
1588 | ||
1589 | tf->device |= ATA_LBA; | |
1590 | } else { | |
1591 | /* CHS */ | |
1592 | u32 sect, head, cyl, track; | |
1593 | ||
1594 | if (!lba_28_ok(block, n_block)) | |
1595 | goto out_of_range; | |
1596 | ||
1597 | /* Convert LBA to CHS */ | |
1598 | track = (u32)block / dev->sectors; | |
1599 | cyl = track / dev->heads; | |
1600 | head = track % dev->heads; | |
1601 | sect = (u32)block % dev->sectors + 1; | |
1602 | ||
1603 | DPRINTK("block %u track %u cyl %u head %u sect %u\n", | |
1604 | (u32)block, track, cyl, head, sect); | |
1605 | ||
1606 | /* Check whether the converted CHS can fit. | |
1607 | Cylinder: 0-65535 | |
1608 | Head: 0-15 | |
1609 | Sector: 1-255*/ | |
1610 | if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect)) | |
1611 | goto out_of_range; | |
1612 | ||
1613 | tf->command = ATA_CMD_VERIFY; | |
1614 | tf->nsect = n_block & 0xff; /* Sector count 0 means 256 sectors */ | |
1615 | tf->lbal = sect; | |
1616 | tf->lbam = cyl; | |
1617 | tf->lbah = cyl >> 8; | |
1618 | tf->device |= head; | |
1619 | } | |
1620 | ||
1621 | return 0; | |
1622 | ||
1623 | invalid_fld: | |
1624 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
1625 | /* "Invalid field in cbd" */ | |
1626 | return 1; | |
1627 | ||
1628 | out_of_range: | |
1629 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); | |
1630 | /* "Logical Block Address out of range" */ | |
1631 | return 1; | |
1632 | ||
1633 | nothing_to_do: | |
1634 | scmd->result = SAM_STAT_GOOD; | |
1635 | return 1; | |
1636 | } | |
1637 | ||
1638 | /** | |
1639 | * ata_scsi_rw_xlat - Translate SCSI r/w command into an ATA one | |
1640 | * @qc: Storage for translated ATA taskfile | |
1641 | * | |
1642 | * Converts any of six SCSI read/write commands into the | |
1643 | * ATA counterpart, including starting sector (LBA), | |
1644 | * sector count, and taking into account the device's LBA48 | |
1645 | * support. | |
1646 | * | |
1647 | * Commands %READ_6, %READ_10, %READ_16, %WRITE_6, %WRITE_10, and | |
1648 | * %WRITE_16 are currently supported. | |
1649 | * | |
1650 | * LOCKING: | |
1651 | * spin_lock_irqsave(host lock) | |
1652 | * | |
1653 | * RETURNS: | |
1654 | * Zero on success, non-zero on error. | |
1655 | */ | |
1656 | static unsigned int ata_scsi_rw_xlat(struct ata_queued_cmd *qc) | |
1657 | { | |
1658 | struct scsi_cmnd *scmd = qc->scsicmd; | |
1659 | const u8 *cdb = scmd->cmnd; | |
1660 | unsigned int tf_flags = 0; | |
1661 | u64 block; | |
1662 | u32 n_block; | |
1663 | int rc; | |
1664 | ||
1665 | if (cdb[0] == WRITE_10 || cdb[0] == WRITE_6 || cdb[0] == WRITE_16) | |
1666 | tf_flags |= ATA_TFLAG_WRITE; | |
1667 | ||
1668 | /* Calculate the SCSI LBA, transfer length and FUA. */ | |
1669 | switch (cdb[0]) { | |
1670 | case READ_10: | |
1671 | case WRITE_10: | |
1672 | if (unlikely(scmd->cmd_len < 10)) | |
1673 | goto invalid_fld; | |
1674 | scsi_10_lba_len(cdb, &block, &n_block); | |
1675 | if (cdb[1] & (1 << 3)) | |
1676 | tf_flags |= ATA_TFLAG_FUA; | |
1677 | break; | |
1678 | case READ_6: | |
1679 | case WRITE_6: | |
1680 | if (unlikely(scmd->cmd_len < 6)) | |
1681 | goto invalid_fld; | |
1682 | scsi_6_lba_len(cdb, &block, &n_block); | |
1683 | ||
1684 | /* for 6-byte r/w commands, transfer length 0 | |
1685 | * means 256 blocks of data, not 0 block. | |
1686 | */ | |
1687 | if (!n_block) | |
1688 | n_block = 256; | |
1689 | break; | |
1690 | case READ_16: | |
1691 | case WRITE_16: | |
1692 | if (unlikely(scmd->cmd_len < 16)) | |
1693 | goto invalid_fld; | |
1694 | scsi_16_lba_len(cdb, &block, &n_block); | |
1695 | if (cdb[1] & (1 << 3)) | |
1696 | tf_flags |= ATA_TFLAG_FUA; | |
1697 | break; | |
1698 | default: | |
1699 | DPRINTK("no-byte command\n"); | |
1700 | goto invalid_fld; | |
1701 | } | |
1702 | ||
1703 | /* Check and compose ATA command */ | |
1704 | if (!n_block) | |
1705 | /* For 10-byte and 16-byte SCSI R/W commands, transfer | |
1706 | * length 0 means transfer 0 block of data. | |
1707 | * However, for ATA R/W commands, sector count 0 means | |
1708 | * 256 or 65536 sectors, not 0 sectors as in SCSI. | |
1709 | * | |
1710 | * WARNING: one or two older ATA drives treat 0 as 0... | |
1711 | */ | |
1712 | goto nothing_to_do; | |
1713 | ||
1714 | qc->flags |= ATA_QCFLAG_IO; | |
1715 | qc->nbytes = n_block * scmd->device->sector_size; | |
1716 | ||
1717 | rc = ata_build_rw_tf(&qc->tf, qc->dev, block, n_block, tf_flags, | |
1718 | qc->tag); | |
1719 | if (likely(rc == 0)) | |
1720 | return 0; | |
1721 | ||
1722 | if (rc == -ERANGE) | |
1723 | goto out_of_range; | |
1724 | /* treat all other errors as -EINVAL, fall through */ | |
1725 | invalid_fld: | |
1726 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
1727 | /* "Invalid field in cbd" */ | |
1728 | return 1; | |
1729 | ||
1730 | out_of_range: | |
1731 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); | |
1732 | /* "Logical Block Address out of range" */ | |
1733 | return 1; | |
1734 | ||
1735 | nothing_to_do: | |
1736 | scmd->result = SAM_STAT_GOOD; | |
1737 | return 1; | |
1738 | } | |
1739 | ||
1740 | static void ata_scsi_qc_complete(struct ata_queued_cmd *qc) | |
1741 | { | |
1742 | struct ata_port *ap = qc->ap; | |
1743 | struct scsi_cmnd *cmd = qc->scsicmd; | |
1744 | u8 *cdb = cmd->cmnd; | |
1745 | int need_sense = (qc->err_mask != 0); | |
1746 | ||
1747 | /* For ATA pass thru (SAT) commands, generate a sense block if | |
1748 | * user mandated it or if there's an error. Note that if we | |
1749 | * generate because the user forced us to [CK_COND =1], a check | |
1750 | * condition is generated and the ATA register values are returned | |
1751 | * whether the command completed successfully or not. If there | |
1752 | * was no error, we use the following sense data: | |
1753 | * sk = RECOVERED ERROR | |
1754 | * asc,ascq = ATA PASS-THROUGH INFORMATION AVAILABLE | |
1755 | */ | |
1756 | if (((cdb[0] == ATA_16) || (cdb[0] == ATA_12)) && | |
1757 | ((cdb[2] & 0x20) || need_sense)) { | |
1758 | ata_gen_passthru_sense(qc); | |
1759 | } else { | |
1760 | if (!need_sense) { | |
1761 | cmd->result = SAM_STAT_GOOD; | |
1762 | } else { | |
1763 | /* TODO: decide which descriptor format to use | |
1764 | * for 48b LBA devices and call that here | |
1765 | * instead of the fixed desc, which is only | |
1766 | * good for smaller LBA (and maybe CHS?) | |
1767 | * devices. | |
1768 | */ | |
1769 | ata_gen_ata_sense(qc); | |
1770 | } | |
1771 | } | |
1772 | ||
1773 | if (need_sense && !ap->ops->error_handler) | |
1774 | ata_dump_status(ap->print_id, &qc->result_tf); | |
1775 | ||
1776 | qc->scsidone(cmd); | |
1777 | ||
1778 | ata_qc_free(qc); | |
1779 | } | |
1780 | ||
1781 | /** | |
1782 | * ata_scsi_translate - Translate then issue SCSI command to ATA device | |
1783 | * @dev: ATA device to which the command is addressed | |
1784 | * @cmd: SCSI command to execute | |
1785 | * @xlat_func: Actor which translates @cmd to an ATA taskfile | |
1786 | * | |
1787 | * Our ->queuecommand() function has decided that the SCSI | |
1788 | * command issued can be directly translated into an ATA | |
1789 | * command, rather than handled internally. | |
1790 | * | |
1791 | * This function sets up an ata_queued_cmd structure for the | |
1792 | * SCSI command, and sends that ata_queued_cmd to the hardware. | |
1793 | * | |
1794 | * The xlat_func argument (actor) returns 0 if ready to execute | |
1795 | * ATA command, else 1 to finish translation. If 1 is returned | |
1796 | * then cmd->result (and possibly cmd->sense_buffer) are assumed | |
1797 | * to be set reflecting an error condition or clean (early) | |
1798 | * termination. | |
1799 | * | |
1800 | * LOCKING: | |
1801 | * spin_lock_irqsave(host lock) | |
1802 | * | |
1803 | * RETURNS: | |
1804 | * 0 on success, SCSI_ML_QUEUE_DEVICE_BUSY if the command | |
1805 | * needs to be deferred. | |
1806 | */ | |
1807 | static int ata_scsi_translate(struct ata_device *dev, struct scsi_cmnd *cmd, | |
1808 | ata_xlat_func_t xlat_func) | |
1809 | { | |
1810 | struct ata_port *ap = dev->link->ap; | |
1811 | struct ata_queued_cmd *qc; | |
1812 | int rc; | |
1813 | ||
1814 | VPRINTK("ENTER\n"); | |
1815 | ||
1816 | qc = ata_scsi_qc_new(dev, cmd); | |
1817 | if (!qc) | |
1818 | goto err_mem; | |
1819 | ||
1820 | /* data is present; dma-map it */ | |
1821 | if (cmd->sc_data_direction == DMA_FROM_DEVICE || | |
1822 | cmd->sc_data_direction == DMA_TO_DEVICE) { | |
1823 | if (unlikely(scsi_bufflen(cmd) < 1)) { | |
1824 | ata_dev_warn(dev, "WARNING: zero len r/w req\n"); | |
1825 | goto err_did; | |
1826 | } | |
1827 | ||
1828 | ata_sg_init(qc, scsi_sglist(cmd), scsi_sg_count(cmd)); | |
1829 | ||
1830 | qc->dma_dir = cmd->sc_data_direction; | |
1831 | } | |
1832 | ||
1833 | qc->complete_fn = ata_scsi_qc_complete; | |
1834 | ||
1835 | if (xlat_func(qc)) | |
1836 | goto early_finish; | |
1837 | ||
1838 | if (ap->ops->qc_defer) { | |
1839 | if ((rc = ap->ops->qc_defer(qc))) | |
1840 | goto defer; | |
1841 | } | |
1842 | ||
1843 | /* select device, send command to hardware */ | |
1844 | ata_qc_issue(qc); | |
1845 | ||
1846 | VPRINTK("EXIT\n"); | |
1847 | return 0; | |
1848 | ||
1849 | early_finish: | |
1850 | ata_qc_free(qc); | |
1851 | cmd->scsi_done(cmd); | |
1852 | DPRINTK("EXIT - early finish (good or error)\n"); | |
1853 | return 0; | |
1854 | ||
1855 | err_did: | |
1856 | ata_qc_free(qc); | |
1857 | cmd->result = (DID_ERROR << 16); | |
1858 | cmd->scsi_done(cmd); | |
1859 | err_mem: | |
1860 | DPRINTK("EXIT - internal\n"); | |
1861 | return 0; | |
1862 | ||
1863 | defer: | |
1864 | ata_qc_free(qc); | |
1865 | DPRINTK("EXIT - defer\n"); | |
1866 | if (rc == ATA_DEFER_LINK) | |
1867 | return SCSI_MLQUEUE_DEVICE_BUSY; | |
1868 | else | |
1869 | return SCSI_MLQUEUE_HOST_BUSY; | |
1870 | } | |
1871 | ||
1872 | /** | |
1873 | * ata_scsi_rbuf_get - Map response buffer. | |
1874 | * @cmd: SCSI command containing buffer to be mapped. | |
1875 | * @flags: unsigned long variable to store irq enable status | |
1876 | * @copy_in: copy in from user buffer | |
1877 | * | |
1878 | * Prepare buffer for simulated SCSI commands. | |
1879 | * | |
1880 | * LOCKING: | |
1881 | * spin_lock_irqsave(ata_scsi_rbuf_lock) on success | |
1882 | * | |
1883 | * RETURNS: | |
1884 | * Pointer to response buffer. | |
1885 | */ | |
1886 | static void *ata_scsi_rbuf_get(struct scsi_cmnd *cmd, bool copy_in, | |
1887 | unsigned long *flags) | |
1888 | { | |
1889 | spin_lock_irqsave(&ata_scsi_rbuf_lock, *flags); | |
1890 | ||
1891 | memset(ata_scsi_rbuf, 0, ATA_SCSI_RBUF_SIZE); | |
1892 | if (copy_in) | |
1893 | sg_copy_to_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), | |
1894 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); | |
1895 | return ata_scsi_rbuf; | |
1896 | } | |
1897 | ||
1898 | /** | |
1899 | * ata_scsi_rbuf_put - Unmap response buffer. | |
1900 | * @cmd: SCSI command containing buffer to be unmapped. | |
1901 | * @copy_out: copy out result | |
1902 | * @flags: @flags passed to ata_scsi_rbuf_get() | |
1903 | * | |
1904 | * Returns rbuf buffer. The result is copied to @cmd's buffer if | |
1905 | * @copy_back is true. | |
1906 | * | |
1907 | * LOCKING: | |
1908 | * Unlocks ata_scsi_rbuf_lock. | |
1909 | */ | |
1910 | static inline void ata_scsi_rbuf_put(struct scsi_cmnd *cmd, bool copy_out, | |
1911 | unsigned long *flags) | |
1912 | { | |
1913 | if (copy_out) | |
1914 | sg_copy_from_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), | |
1915 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); | |
1916 | spin_unlock_irqrestore(&ata_scsi_rbuf_lock, *flags); | |
1917 | } | |
1918 | ||
1919 | /** | |
1920 | * ata_scsi_rbuf_fill - wrapper for SCSI command simulators | |
1921 | * @args: device IDENTIFY data / SCSI command of interest. | |
1922 | * @actor: Callback hook for desired SCSI command simulator | |
1923 | * | |
1924 | * Takes care of the hard work of simulating a SCSI command... | |
1925 | * Mapping the response buffer, calling the command's handler, | |
1926 | * and handling the handler's return value. This return value | |
1927 | * indicates whether the handler wishes the SCSI command to be | |
1928 | * completed successfully (0), or not (in which case cmd->result | |
1929 | * and sense buffer are assumed to be set). | |
1930 | * | |
1931 | * LOCKING: | |
1932 | * spin_lock_irqsave(host lock) | |
1933 | */ | |
1934 | static void ata_scsi_rbuf_fill(struct ata_scsi_args *args, | |
1935 | unsigned int (*actor)(struct ata_scsi_args *args, u8 *rbuf)) | |
1936 | { | |
1937 | u8 *rbuf; | |
1938 | unsigned int rc; | |
1939 | struct scsi_cmnd *cmd = args->cmd; | |
1940 | unsigned long flags; | |
1941 | ||
1942 | rbuf = ata_scsi_rbuf_get(cmd, false, &flags); | |
1943 | rc = actor(args, rbuf); | |
1944 | ata_scsi_rbuf_put(cmd, rc == 0, &flags); | |
1945 | ||
1946 | if (rc == 0) | |
1947 | cmd->result = SAM_STAT_GOOD; | |
1948 | args->done(cmd); | |
1949 | } | |
1950 | ||
1951 | /** | |
1952 | * ata_scsiop_inq_std - Simulate INQUIRY command | |
1953 | * @args: device IDENTIFY data / SCSI command of interest. | |
1954 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
1955 | * | |
1956 | * Returns standard device identification data associated | |
1957 | * with non-VPD INQUIRY command output. | |
1958 | * | |
1959 | * LOCKING: | |
1960 | * spin_lock_irqsave(host lock) | |
1961 | */ | |
1962 | static unsigned int ata_scsiop_inq_std(struct ata_scsi_args *args, u8 *rbuf) | |
1963 | { | |
1964 | const u8 versions[] = { | |
1965 | 0x00, | |
1966 | 0x60, /* SAM-3 (no version claimed) */ | |
1967 | ||
1968 | 0x03, | |
1969 | 0x20, /* SBC-2 (no version claimed) */ | |
1970 | ||
1971 | 0x02, | |
1972 | 0x60 /* SPC-3 (no version claimed) */ | |
1973 | }; | |
1974 | const u8 versions_zbc[] = { | |
1975 | 0x00, | |
1976 | 0xA0, /* SAM-5 (no version claimed) */ | |
1977 | ||
1978 | 0x04, | |
1979 | 0xC0, /* SBC-3 (no version claimed) */ | |
1980 | ||
1981 | 0x04, | |
1982 | 0x60, /* SPC-4 (no version claimed) */ | |
1983 | ||
1984 | 0x60, | |
1985 | 0x20, /* ZBC (no version claimed) */ | |
1986 | }; | |
1987 | ||
1988 | u8 hdr[] = { | |
1989 | TYPE_DISK, | |
1990 | 0, | |
1991 | 0x5, /* claim SPC-3 version compatibility */ | |
1992 | 2, | |
1993 | 95 - 4 | |
1994 | }; | |
1995 | ||
1996 | VPRINTK("ENTER\n"); | |
1997 | ||
1998 | /* set scsi removeable (RMB) bit per ata bit */ | |
1999 | if (ata_id_removeable(args->id)) | |
2000 | hdr[1] |= (1 << 7); | |
2001 | ||
2002 | if (args->dev->class == ATA_DEV_ZAC) { | |
2003 | hdr[0] = TYPE_ZBC; | |
2004 | hdr[2] = 0x6; /* ZBC is defined in SPC-4 */ | |
2005 | } | |
2006 | ||
2007 | memcpy(rbuf, hdr, sizeof(hdr)); | |
2008 | memcpy(&rbuf[8], "ATA ", 8); | |
2009 | ata_id_string(args->id, &rbuf[16], ATA_ID_PROD, 16); | |
2010 | ||
2011 | /* From SAT, use last 2 words from fw rev unless they are spaces */ | |
2012 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV + 2, 4); | |
2013 | if (strncmp(&rbuf[32], " ", 4) == 0) | |
2014 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV, 4); | |
2015 | ||
2016 | if (rbuf[32] == 0 || rbuf[32] == ' ') | |
2017 | memcpy(&rbuf[32], "n/a ", 4); | |
2018 | ||
2019 | if (args->dev->class == ATA_DEV_ZAC) | |
2020 | memcpy(rbuf + 58, versions_zbc, sizeof(versions_zbc)); | |
2021 | else | |
2022 | memcpy(rbuf + 58, versions, sizeof(versions)); | |
2023 | ||
2024 | return 0; | |
2025 | } | |
2026 | ||
2027 | /** | |
2028 | * ata_scsiop_inq_00 - Simulate INQUIRY VPD page 0, list of pages | |
2029 | * @args: device IDENTIFY data / SCSI command of interest. | |
2030 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2031 | * | |
2032 | * Returns list of inquiry VPD pages available. | |
2033 | * | |
2034 | * LOCKING: | |
2035 | * spin_lock_irqsave(host lock) | |
2036 | */ | |
2037 | static unsigned int ata_scsiop_inq_00(struct ata_scsi_args *args, u8 *rbuf) | |
2038 | { | |
2039 | const u8 pages[] = { | |
2040 | 0x00, /* page 0x00, this page */ | |
2041 | 0x80, /* page 0x80, unit serial no page */ | |
2042 | 0x83, /* page 0x83, device ident page */ | |
2043 | 0x89, /* page 0x89, ata info page */ | |
2044 | 0xb0, /* page 0xb0, block limits page */ | |
2045 | 0xb1, /* page 0xb1, block device characteristics page */ | |
2046 | 0xb2, /* page 0xb2, thin provisioning page */ | |
2047 | }; | |
2048 | ||
2049 | rbuf[3] = sizeof(pages); /* number of supported VPD pages */ | |
2050 | memcpy(rbuf + 4, pages, sizeof(pages)); | |
2051 | return 0; | |
2052 | } | |
2053 | ||
2054 | /** | |
2055 | * ata_scsiop_inq_80 - Simulate INQUIRY VPD page 80, device serial number | |
2056 | * @args: device IDENTIFY data / SCSI command of interest. | |
2057 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2058 | * | |
2059 | * Returns ATA device serial number. | |
2060 | * | |
2061 | * LOCKING: | |
2062 | * spin_lock_irqsave(host lock) | |
2063 | */ | |
2064 | static unsigned int ata_scsiop_inq_80(struct ata_scsi_args *args, u8 *rbuf) | |
2065 | { | |
2066 | const u8 hdr[] = { | |
2067 | 0, | |
2068 | 0x80, /* this page code */ | |
2069 | 0, | |
2070 | ATA_ID_SERNO_LEN, /* page len */ | |
2071 | }; | |
2072 | ||
2073 | memcpy(rbuf, hdr, sizeof(hdr)); | |
2074 | ata_id_string(args->id, (unsigned char *) &rbuf[4], | |
2075 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); | |
2076 | return 0; | |
2077 | } | |
2078 | ||
2079 | /** | |
2080 | * ata_scsiop_inq_83 - Simulate INQUIRY VPD page 83, device identity | |
2081 | * @args: device IDENTIFY data / SCSI command of interest. | |
2082 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2083 | * | |
2084 | * Yields two logical unit device identification designators: | |
2085 | * - vendor specific ASCII containing the ATA serial number | |
2086 | * - SAT defined "t10 vendor id based" containing ASCII vendor | |
2087 | * name ("ATA "), model and serial numbers. | |
2088 | * | |
2089 | * LOCKING: | |
2090 | * spin_lock_irqsave(host lock) | |
2091 | */ | |
2092 | static unsigned int ata_scsiop_inq_83(struct ata_scsi_args *args, u8 *rbuf) | |
2093 | { | |
2094 | const int sat_model_serial_desc_len = 68; | |
2095 | int num; | |
2096 | ||
2097 | rbuf[1] = 0x83; /* this page code */ | |
2098 | num = 4; | |
2099 | ||
2100 | /* piv=0, assoc=lu, code_set=ACSII, designator=vendor */ | |
2101 | rbuf[num + 0] = 2; | |
2102 | rbuf[num + 3] = ATA_ID_SERNO_LEN; | |
2103 | num += 4; | |
2104 | ata_id_string(args->id, (unsigned char *) rbuf + num, | |
2105 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); | |
2106 | num += ATA_ID_SERNO_LEN; | |
2107 | ||
2108 | /* SAT defined lu model and serial numbers descriptor */ | |
2109 | /* piv=0, assoc=lu, code_set=ACSII, designator=t10 vendor id */ | |
2110 | rbuf[num + 0] = 2; | |
2111 | rbuf[num + 1] = 1; | |
2112 | rbuf[num + 3] = sat_model_serial_desc_len; | |
2113 | num += 4; | |
2114 | memcpy(rbuf + num, "ATA ", 8); | |
2115 | num += 8; | |
2116 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_PROD, | |
2117 | ATA_ID_PROD_LEN); | |
2118 | num += ATA_ID_PROD_LEN; | |
2119 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_SERNO, | |
2120 | ATA_ID_SERNO_LEN); | |
2121 | num += ATA_ID_SERNO_LEN; | |
2122 | ||
2123 | if (ata_id_has_wwn(args->id)) { | |
2124 | /* SAT defined lu world wide name */ | |
2125 | /* piv=0, assoc=lu, code_set=binary, designator=NAA */ | |
2126 | rbuf[num + 0] = 1; | |
2127 | rbuf[num + 1] = 3; | |
2128 | rbuf[num + 3] = ATA_ID_WWN_LEN; | |
2129 | num += 4; | |
2130 | ata_id_string(args->id, (unsigned char *) rbuf + num, | |
2131 | ATA_ID_WWN, ATA_ID_WWN_LEN); | |
2132 | num += ATA_ID_WWN_LEN; | |
2133 | } | |
2134 | rbuf[3] = num - 4; /* page len (assume less than 256 bytes) */ | |
2135 | return 0; | |
2136 | } | |
2137 | ||
2138 | /** | |
2139 | * ata_scsiop_inq_89 - Simulate INQUIRY VPD page 89, ATA info | |
2140 | * @args: device IDENTIFY data / SCSI command of interest. | |
2141 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2142 | * | |
2143 | * Yields SAT-specified ATA VPD page. | |
2144 | * | |
2145 | * LOCKING: | |
2146 | * spin_lock_irqsave(host lock) | |
2147 | */ | |
2148 | static unsigned int ata_scsiop_inq_89(struct ata_scsi_args *args, u8 *rbuf) | |
2149 | { | |
2150 | struct ata_taskfile tf; | |
2151 | ||
2152 | memset(&tf, 0, sizeof(tf)); | |
2153 | ||
2154 | rbuf[1] = 0x89; /* our page code */ | |
2155 | rbuf[2] = (0x238 >> 8); /* page size fixed at 238h */ | |
2156 | rbuf[3] = (0x238 & 0xff); | |
2157 | ||
2158 | memcpy(&rbuf[8], "linux ", 8); | |
2159 | memcpy(&rbuf[16], "libata ", 16); | |
2160 | memcpy(&rbuf[32], DRV_VERSION, 4); | |
2161 | ||
2162 | /* we don't store the ATA device signature, so we fake it */ | |
2163 | ||
2164 | tf.command = ATA_DRDY; /* really, this is Status reg */ | |
2165 | tf.lbal = 0x1; | |
2166 | tf.nsect = 0x1; | |
2167 | ||
2168 | ata_tf_to_fis(&tf, 0, 1, &rbuf[36]); /* TODO: PMP? */ | |
2169 | rbuf[36] = 0x34; /* force D2H Reg FIS (34h) */ | |
2170 | ||
2171 | rbuf[56] = ATA_CMD_ID_ATA; | |
2172 | ||
2173 | memcpy(&rbuf[60], &args->id[0], 512); | |
2174 | return 0; | |
2175 | } | |
2176 | ||
2177 | static unsigned int ata_scsiop_inq_b0(struct ata_scsi_args *args, u8 *rbuf) | |
2178 | { | |
2179 | u16 min_io_sectors; | |
2180 | ||
2181 | rbuf[1] = 0xb0; | |
2182 | rbuf[3] = 0x3c; /* required VPD size with unmap support */ | |
2183 | ||
2184 | /* | |
2185 | * Optimal transfer length granularity. | |
2186 | * | |
2187 | * This is always one physical block, but for disks with a smaller | |
2188 | * logical than physical sector size we need to figure out what the | |
2189 | * latter is. | |
2190 | */ | |
2191 | min_io_sectors = 1 << ata_id_log2_per_physical_sector(args->id); | |
2192 | put_unaligned_be16(min_io_sectors, &rbuf[6]); | |
2193 | ||
2194 | /* | |
2195 | * Optimal unmap granularity. | |
2196 | * | |
2197 | * The ATA spec doesn't even know about a granularity or alignment | |
2198 | * for the TRIM command. We can leave away most of the unmap related | |
2199 | * VPD page entries, but we have specifify a granularity to signal | |
2200 | * that we support some form of unmap - in thise case via WRITE SAME | |
2201 | * with the unmap bit set. | |
2202 | */ | |
2203 | if (ata_id_has_trim(args->id)) { | |
2204 | put_unaligned_be64(65535 * 512 / 8, &rbuf[36]); | |
2205 | put_unaligned_be32(1, &rbuf[28]); | |
2206 | } | |
2207 | ||
2208 | return 0; | |
2209 | } | |
2210 | ||
2211 | static unsigned int ata_scsiop_inq_b1(struct ata_scsi_args *args, u8 *rbuf) | |
2212 | { | |
2213 | int form_factor = ata_id_form_factor(args->id); | |
2214 | int media_rotation_rate = ata_id_rotation_rate(args->id); | |
2215 | ||
2216 | rbuf[1] = 0xb1; | |
2217 | rbuf[3] = 0x3c; | |
2218 | rbuf[4] = media_rotation_rate >> 8; | |
2219 | rbuf[5] = media_rotation_rate; | |
2220 | rbuf[7] = form_factor; | |
2221 | ||
2222 | return 0; | |
2223 | } | |
2224 | ||
2225 | static unsigned int ata_scsiop_inq_b2(struct ata_scsi_args *args, u8 *rbuf) | |
2226 | { | |
2227 | /* SCSI Thin Provisioning VPD page: SBC-3 rev 22 or later */ | |
2228 | rbuf[1] = 0xb2; | |
2229 | rbuf[3] = 0x4; | |
2230 | rbuf[5] = 1 << 6; /* TPWS */ | |
2231 | ||
2232 | return 0; | |
2233 | } | |
2234 | ||
2235 | /** | |
2236 | * ata_scsiop_noop - Command handler that simply returns success. | |
2237 | * @args: device IDENTIFY data / SCSI command of interest. | |
2238 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2239 | * | |
2240 | * No operation. Simply returns success to caller, to indicate | |
2241 | * that the caller should successfully complete this SCSI command. | |
2242 | * | |
2243 | * LOCKING: | |
2244 | * spin_lock_irqsave(host lock) | |
2245 | */ | |
2246 | static unsigned int ata_scsiop_noop(struct ata_scsi_args *args, u8 *rbuf) | |
2247 | { | |
2248 | VPRINTK("ENTER\n"); | |
2249 | return 0; | |
2250 | } | |
2251 | ||
2252 | /** | |
2253 | * modecpy - Prepare response for MODE SENSE | |
2254 | * @dest: output buffer | |
2255 | * @src: data being copied | |
2256 | * @n: length of mode page | |
2257 | * @changeable: whether changeable parameters are requested | |
2258 | * | |
2259 | * Generate a generic MODE SENSE page for either current or changeable | |
2260 | * parameters. | |
2261 | * | |
2262 | * LOCKING: | |
2263 | * None. | |
2264 | */ | |
2265 | static void modecpy(u8 *dest, const u8 *src, int n, bool changeable) | |
2266 | { | |
2267 | if (changeable) { | |
2268 | memcpy(dest, src, 2); | |
2269 | memset(dest + 2, 0, n - 2); | |
2270 | } else { | |
2271 | memcpy(dest, src, n); | |
2272 | } | |
2273 | } | |
2274 | ||
2275 | /** | |
2276 | * ata_msense_caching - Simulate MODE SENSE caching info page | |
2277 | * @id: device IDENTIFY data | |
2278 | * @buf: output buffer | |
2279 | * @changeable: whether changeable parameters are requested | |
2280 | * | |
2281 | * Generate a caching info page, which conditionally indicates | |
2282 | * write caching to the SCSI layer, depending on device | |
2283 | * capabilities. | |
2284 | * | |
2285 | * LOCKING: | |
2286 | * None. | |
2287 | */ | |
2288 | static unsigned int ata_msense_caching(u16 *id, u8 *buf, bool changeable) | |
2289 | { | |
2290 | modecpy(buf, def_cache_mpage, sizeof(def_cache_mpage), changeable); | |
2291 | if (changeable || ata_id_wcache_enabled(id)) | |
2292 | buf[2] |= (1 << 2); /* write cache enable */ | |
2293 | if (!changeable && !ata_id_rahead_enabled(id)) | |
2294 | buf[12] |= (1 << 5); /* disable read ahead */ | |
2295 | return sizeof(def_cache_mpage); | |
2296 | } | |
2297 | ||
2298 | /** | |
2299 | * ata_msense_ctl_mode - Simulate MODE SENSE control mode page | |
2300 | * @buf: output buffer | |
2301 | * @changeable: whether changeable parameters are requested | |
2302 | * | |
2303 | * Generate a generic MODE SENSE control mode page. | |
2304 | * | |
2305 | * LOCKING: | |
2306 | * None. | |
2307 | */ | |
2308 | static unsigned int ata_msense_ctl_mode(u8 *buf, bool changeable) | |
2309 | { | |
2310 | modecpy(buf, def_control_mpage, sizeof(def_control_mpage), changeable); | |
2311 | return sizeof(def_control_mpage); | |
2312 | } | |
2313 | ||
2314 | /** | |
2315 | * ata_msense_rw_recovery - Simulate MODE SENSE r/w error recovery page | |
2316 | * @buf: output buffer | |
2317 | * @changeable: whether changeable parameters are requested | |
2318 | * | |
2319 | * Generate a generic MODE SENSE r/w error recovery page. | |
2320 | * | |
2321 | * LOCKING: | |
2322 | * None. | |
2323 | */ | |
2324 | static unsigned int ata_msense_rw_recovery(u8 *buf, bool changeable) | |
2325 | { | |
2326 | modecpy(buf, def_rw_recovery_mpage, sizeof(def_rw_recovery_mpage), | |
2327 | changeable); | |
2328 | return sizeof(def_rw_recovery_mpage); | |
2329 | } | |
2330 | ||
2331 | /* | |
2332 | * We can turn this into a real blacklist if it's needed, for now just | |
2333 | * blacklist any Maxtor BANC1G10 revision firmware | |
2334 | */ | |
2335 | static int ata_dev_supports_fua(u16 *id) | |
2336 | { | |
2337 | unsigned char model[ATA_ID_PROD_LEN + 1], fw[ATA_ID_FW_REV_LEN + 1]; | |
2338 | ||
2339 | if (!libata_fua) | |
2340 | return 0; | |
2341 | if (!ata_id_has_fua(id)) | |
2342 | return 0; | |
2343 | ||
2344 | ata_id_c_string(id, model, ATA_ID_PROD, sizeof(model)); | |
2345 | ata_id_c_string(id, fw, ATA_ID_FW_REV, sizeof(fw)); | |
2346 | ||
2347 | if (strcmp(model, "Maxtor")) | |
2348 | return 1; | |
2349 | if (strcmp(fw, "BANC1G10")) | |
2350 | return 1; | |
2351 | ||
2352 | return 0; /* blacklisted */ | |
2353 | } | |
2354 | ||
2355 | /** | |
2356 | * ata_scsiop_mode_sense - Simulate MODE SENSE 6, 10 commands | |
2357 | * @args: device IDENTIFY data / SCSI command of interest. | |
2358 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2359 | * | |
2360 | * Simulate MODE SENSE commands. Assume this is invoked for direct | |
2361 | * access devices (e.g. disks) only. There should be no block | |
2362 | * descriptor for other device types. | |
2363 | * | |
2364 | * LOCKING: | |
2365 | * spin_lock_irqsave(host lock) | |
2366 | */ | |
2367 | static unsigned int ata_scsiop_mode_sense(struct ata_scsi_args *args, u8 *rbuf) | |
2368 | { | |
2369 | struct ata_device *dev = args->dev; | |
2370 | u8 *scsicmd = args->cmd->cmnd, *p = rbuf; | |
2371 | const u8 sat_blk_desc[] = { | |
2372 | 0, 0, 0, 0, /* number of blocks: sat unspecified */ | |
2373 | 0, | |
2374 | 0, 0x2, 0x0 /* block length: 512 bytes */ | |
2375 | }; | |
2376 | u8 pg, spg; | |
2377 | unsigned int ebd, page_control, six_byte; | |
2378 | u8 dpofua; | |
2379 | ||
2380 | VPRINTK("ENTER\n"); | |
2381 | ||
2382 | six_byte = (scsicmd[0] == MODE_SENSE); | |
2383 | ebd = !(scsicmd[1] & 0x8); /* dbd bit inverted == edb */ | |
2384 | /* | |
2385 | * LLBA bit in msense(10) ignored (compliant) | |
2386 | */ | |
2387 | ||
2388 | page_control = scsicmd[2] >> 6; | |
2389 | switch (page_control) { | |
2390 | case 0: /* current */ | |
2391 | case 1: /* changeable */ | |
2392 | case 2: /* defaults */ | |
2393 | break; /* supported */ | |
2394 | case 3: /* saved */ | |
2395 | goto saving_not_supp; | |
2396 | default: | |
2397 | goto invalid_fld; | |
2398 | } | |
2399 | ||
2400 | if (six_byte) | |
2401 | p += 4 + (ebd ? 8 : 0); | |
2402 | else | |
2403 | p += 8 + (ebd ? 8 : 0); | |
2404 | ||
2405 | pg = scsicmd[2] & 0x3f; | |
2406 | spg = scsicmd[3]; | |
2407 | /* | |
2408 | * No mode subpages supported (yet) but asking for _all_ | |
2409 | * subpages may be valid | |
2410 | */ | |
2411 | if (spg && (spg != ALL_SUB_MPAGES)) | |
2412 | goto invalid_fld; | |
2413 | ||
2414 | switch(pg) { | |
2415 | case RW_RECOVERY_MPAGE: | |
2416 | p += ata_msense_rw_recovery(p, page_control == 1); | |
2417 | break; | |
2418 | ||
2419 | case CACHE_MPAGE: | |
2420 | p += ata_msense_caching(args->id, p, page_control == 1); | |
2421 | break; | |
2422 | ||
2423 | case CONTROL_MPAGE: | |
2424 | p += ata_msense_ctl_mode(p, page_control == 1); | |
2425 | break; | |
2426 | ||
2427 | case ALL_MPAGES: | |
2428 | p += ata_msense_rw_recovery(p, page_control == 1); | |
2429 | p += ata_msense_caching(args->id, p, page_control == 1); | |
2430 | p += ata_msense_ctl_mode(p, page_control == 1); | |
2431 | break; | |
2432 | ||
2433 | default: /* invalid page code */ | |
2434 | goto invalid_fld; | |
2435 | } | |
2436 | ||
2437 | dpofua = 0; | |
2438 | if (ata_dev_supports_fua(args->id) && (dev->flags & ATA_DFLAG_LBA48) && | |
2439 | (!(dev->flags & ATA_DFLAG_PIO) || dev->multi_count)) | |
2440 | dpofua = 1 << 4; | |
2441 | ||
2442 | if (six_byte) { | |
2443 | rbuf[0] = p - rbuf - 1; | |
2444 | rbuf[2] |= dpofua; | |
2445 | if (ebd) { | |
2446 | rbuf[3] = sizeof(sat_blk_desc); | |
2447 | memcpy(rbuf + 4, sat_blk_desc, sizeof(sat_blk_desc)); | |
2448 | } | |
2449 | } else { | |
2450 | unsigned int output_len = p - rbuf - 2; | |
2451 | ||
2452 | rbuf[0] = output_len >> 8; | |
2453 | rbuf[1] = output_len; | |
2454 | rbuf[3] |= dpofua; | |
2455 | if (ebd) { | |
2456 | rbuf[7] = sizeof(sat_blk_desc); | |
2457 | memcpy(rbuf + 8, sat_blk_desc, sizeof(sat_blk_desc)); | |
2458 | } | |
2459 | } | |
2460 | return 0; | |
2461 | ||
2462 | invalid_fld: | |
2463 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
2464 | /* "Invalid field in cbd" */ | |
2465 | return 1; | |
2466 | ||
2467 | saving_not_supp: | |
2468 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x39, 0x0); | |
2469 | /* "Saving parameters not supported" */ | |
2470 | return 1; | |
2471 | } | |
2472 | ||
2473 | /** | |
2474 | * ata_scsiop_read_cap - Simulate READ CAPACITY[ 16] commands | |
2475 | * @args: device IDENTIFY data / SCSI command of interest. | |
2476 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2477 | * | |
2478 | * Simulate READ CAPACITY commands. | |
2479 | * | |
2480 | * LOCKING: | |
2481 | * None. | |
2482 | */ | |
2483 | static unsigned int ata_scsiop_read_cap(struct ata_scsi_args *args, u8 *rbuf) | |
2484 | { | |
2485 | struct ata_device *dev = args->dev; | |
2486 | u64 last_lba = dev->n_sectors - 1; /* LBA of the last block */ | |
2487 | u32 sector_size; /* physical sector size in bytes */ | |
2488 | u8 log2_per_phys; | |
2489 | u16 lowest_aligned; | |
2490 | ||
2491 | sector_size = ata_id_logical_sector_size(dev->id); | |
2492 | log2_per_phys = ata_id_log2_per_physical_sector(dev->id); | |
2493 | lowest_aligned = ata_id_logical_sector_offset(dev->id, log2_per_phys); | |
2494 | ||
2495 | VPRINTK("ENTER\n"); | |
2496 | ||
2497 | if (args->cmd->cmnd[0] == READ_CAPACITY) { | |
2498 | if (last_lba >= 0xffffffffULL) | |
2499 | last_lba = 0xffffffff; | |
2500 | ||
2501 | /* sector count, 32-bit */ | |
2502 | rbuf[0] = last_lba >> (8 * 3); | |
2503 | rbuf[1] = last_lba >> (8 * 2); | |
2504 | rbuf[2] = last_lba >> (8 * 1); | |
2505 | rbuf[3] = last_lba; | |
2506 | ||
2507 | /* sector size */ | |
2508 | rbuf[4] = sector_size >> (8 * 3); | |
2509 | rbuf[5] = sector_size >> (8 * 2); | |
2510 | rbuf[6] = sector_size >> (8 * 1); | |
2511 | rbuf[7] = sector_size; | |
2512 | } else { | |
2513 | /* sector count, 64-bit */ | |
2514 | rbuf[0] = last_lba >> (8 * 7); | |
2515 | rbuf[1] = last_lba >> (8 * 6); | |
2516 | rbuf[2] = last_lba >> (8 * 5); | |
2517 | rbuf[3] = last_lba >> (8 * 4); | |
2518 | rbuf[4] = last_lba >> (8 * 3); | |
2519 | rbuf[5] = last_lba >> (8 * 2); | |
2520 | rbuf[6] = last_lba >> (8 * 1); | |
2521 | rbuf[7] = last_lba; | |
2522 | ||
2523 | /* sector size */ | |
2524 | rbuf[ 8] = sector_size >> (8 * 3); | |
2525 | rbuf[ 9] = sector_size >> (8 * 2); | |
2526 | rbuf[10] = sector_size >> (8 * 1); | |
2527 | rbuf[11] = sector_size; | |
2528 | ||
2529 | rbuf[12] = 0; | |
2530 | rbuf[13] = log2_per_phys; | |
2531 | rbuf[14] = (lowest_aligned >> 8) & 0x3f; | |
2532 | rbuf[15] = lowest_aligned; | |
2533 | ||
2534 | if (ata_id_has_trim(args->id)) { | |
2535 | rbuf[14] |= 0x80; /* TPE */ | |
2536 | ||
2537 | if (ata_id_has_zero_after_trim(args->id)) | |
2538 | rbuf[14] |= 0x40; /* TPRZ */ | |
2539 | } | |
2540 | } | |
2541 | ||
2542 | return 0; | |
2543 | } | |
2544 | ||
2545 | /** | |
2546 | * ata_scsiop_report_luns - Simulate REPORT LUNS command | |
2547 | * @args: device IDENTIFY data / SCSI command of interest. | |
2548 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. | |
2549 | * | |
2550 | * Simulate REPORT LUNS command. | |
2551 | * | |
2552 | * LOCKING: | |
2553 | * spin_lock_irqsave(host lock) | |
2554 | */ | |
2555 | static unsigned int ata_scsiop_report_luns(struct ata_scsi_args *args, u8 *rbuf) | |
2556 | { | |
2557 | VPRINTK("ENTER\n"); | |
2558 | rbuf[3] = 8; /* just one lun, LUN 0, size 8 bytes */ | |
2559 | ||
2560 | return 0; | |
2561 | } | |
2562 | ||
2563 | static void atapi_sense_complete(struct ata_queued_cmd *qc) | |
2564 | { | |
2565 | if (qc->err_mask && ((qc->err_mask & AC_ERR_DEV) == 0)) { | |
2566 | /* FIXME: not quite right; we don't want the | |
2567 | * translation of taskfile registers into | |
2568 | * a sense descriptors, since that's only | |
2569 | * correct for ATA, not ATAPI | |
2570 | */ | |
2571 | ata_gen_passthru_sense(qc); | |
2572 | } | |
2573 | ||
2574 | qc->scsidone(qc->scsicmd); | |
2575 | ata_qc_free(qc); | |
2576 | } | |
2577 | ||
2578 | /* is it pointless to prefer PIO for "safety reasons"? */ | |
2579 | static inline int ata_pio_use_silly(struct ata_port *ap) | |
2580 | { | |
2581 | return (ap->flags & ATA_FLAG_PIO_DMA); | |
2582 | } | |
2583 | ||
2584 | static void atapi_request_sense(struct ata_queued_cmd *qc) | |
2585 | { | |
2586 | struct ata_port *ap = qc->ap; | |
2587 | struct scsi_cmnd *cmd = qc->scsicmd; | |
2588 | ||
2589 | DPRINTK("ATAPI request sense\n"); | |
2590 | ||
2591 | memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); | |
2592 | ||
2593 | #ifdef CONFIG_ATA_SFF | |
2594 | if (ap->ops->sff_tf_read) | |
2595 | ap->ops->sff_tf_read(ap, &qc->tf); | |
2596 | #endif | |
2597 | ||
2598 | /* fill these in, for the case where they are -not- overwritten */ | |
2599 | cmd->sense_buffer[0] = 0x70; | |
2600 | cmd->sense_buffer[2] = qc->tf.feature >> 4; | |
2601 | ||
2602 | ata_qc_reinit(qc); | |
2603 | ||
2604 | /* setup sg table and init transfer direction */ | |
2605 | sg_init_one(&qc->sgent, cmd->sense_buffer, SCSI_SENSE_BUFFERSIZE); | |
2606 | ata_sg_init(qc, &qc->sgent, 1); | |
2607 | qc->dma_dir = DMA_FROM_DEVICE; | |
2608 | ||
2609 | memset(&qc->cdb, 0, qc->dev->cdb_len); | |
2610 | qc->cdb[0] = REQUEST_SENSE; | |
2611 | qc->cdb[4] = SCSI_SENSE_BUFFERSIZE; | |
2612 | ||
2613 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; | |
2614 | qc->tf.command = ATA_CMD_PACKET; | |
2615 | ||
2616 | if (ata_pio_use_silly(ap)) { | |
2617 | qc->tf.protocol = ATAPI_PROT_DMA; | |
2618 | qc->tf.feature |= ATAPI_PKT_DMA; | |
2619 | } else { | |
2620 | qc->tf.protocol = ATAPI_PROT_PIO; | |
2621 | qc->tf.lbam = SCSI_SENSE_BUFFERSIZE; | |
2622 | qc->tf.lbah = 0; | |
2623 | } | |
2624 | qc->nbytes = SCSI_SENSE_BUFFERSIZE; | |
2625 | ||
2626 | qc->complete_fn = atapi_sense_complete; | |
2627 | ||
2628 | ata_qc_issue(qc); | |
2629 | ||
2630 | DPRINTK("EXIT\n"); | |
2631 | } | |
2632 | ||
2633 | static void atapi_qc_complete(struct ata_queued_cmd *qc) | |
2634 | { | |
2635 | struct scsi_cmnd *cmd = qc->scsicmd; | |
2636 | unsigned int err_mask = qc->err_mask; | |
2637 | ||
2638 | VPRINTK("ENTER, err_mask 0x%X\n", err_mask); | |
2639 | ||
2640 | /* handle completion from new EH */ | |
2641 | if (unlikely(qc->ap->ops->error_handler && | |
2642 | (err_mask || qc->flags & ATA_QCFLAG_SENSE_VALID))) { | |
2643 | ||
2644 | if (!(qc->flags & ATA_QCFLAG_SENSE_VALID)) { | |
2645 | /* FIXME: not quite right; we don't want the | |
2646 | * translation of taskfile registers into a | |
2647 | * sense descriptors, since that's only | |
2648 | * correct for ATA, not ATAPI | |
2649 | */ | |
2650 | ata_gen_passthru_sense(qc); | |
2651 | } | |
2652 | ||
2653 | /* SCSI EH automatically locks door if sdev->locked is | |
2654 | * set. Sometimes door lock request continues to | |
2655 | * fail, for example, when no media is present. This | |
2656 | * creates a loop - SCSI EH issues door lock which | |
2657 | * fails and gets invoked again to acquire sense data | |
2658 | * for the failed command. | |
2659 | * | |
2660 | * If door lock fails, always clear sdev->locked to | |
2661 | * avoid this infinite loop. | |
2662 | * | |
2663 | * This may happen before SCSI scan is complete. Make | |
2664 | * sure qc->dev->sdev isn't NULL before dereferencing. | |
2665 | */ | |
2666 | if (qc->cdb[0] == ALLOW_MEDIUM_REMOVAL && qc->dev->sdev) | |
2667 | qc->dev->sdev->locked = 0; | |
2668 | ||
2669 | qc->scsicmd->result = SAM_STAT_CHECK_CONDITION; | |
2670 | qc->scsidone(cmd); | |
2671 | ata_qc_free(qc); | |
2672 | return; | |
2673 | } | |
2674 | ||
2675 | /* successful completion or old EH failure path */ | |
2676 | if (unlikely(err_mask & AC_ERR_DEV)) { | |
2677 | cmd->result = SAM_STAT_CHECK_CONDITION; | |
2678 | atapi_request_sense(qc); | |
2679 | return; | |
2680 | } else if (unlikely(err_mask)) { | |
2681 | /* FIXME: not quite right; we don't want the | |
2682 | * translation of taskfile registers into | |
2683 | * a sense descriptors, since that's only | |
2684 | * correct for ATA, not ATAPI | |
2685 | */ | |
2686 | ata_gen_passthru_sense(qc); | |
2687 | } else { | |
2688 | u8 *scsicmd = cmd->cmnd; | |
2689 | ||
2690 | if ((scsicmd[0] == INQUIRY) && ((scsicmd[1] & 0x03) == 0)) { | |
2691 | unsigned long flags; | |
2692 | u8 *buf; | |
2693 | ||
2694 | buf = ata_scsi_rbuf_get(cmd, true, &flags); | |
2695 | ||
2696 | /* ATAPI devices typically report zero for their SCSI version, | |
2697 | * and sometimes deviate from the spec WRT response data | |
2698 | * format. If SCSI version is reported as zero like normal, | |
2699 | * then we make the following fixups: 1) Fake MMC-5 version, | |
2700 | * to indicate to the Linux scsi midlayer this is a modern | |
2701 | * device. 2) Ensure response data format / ATAPI information | |
2702 | * are always correct. | |
2703 | */ | |
2704 | if (buf[2] == 0) { | |
2705 | buf[2] = 0x5; | |
2706 | buf[3] = 0x32; | |
2707 | } | |
2708 | ||
2709 | ata_scsi_rbuf_put(cmd, true, &flags); | |
2710 | } | |
2711 | ||
2712 | cmd->result = SAM_STAT_GOOD; | |
2713 | } | |
2714 | ||
2715 | qc->scsidone(cmd); | |
2716 | ata_qc_free(qc); | |
2717 | } | |
2718 | /** | |
2719 | * atapi_xlat - Initialize PACKET taskfile | |
2720 | * @qc: command structure to be initialized | |
2721 | * | |
2722 | * LOCKING: | |
2723 | * spin_lock_irqsave(host lock) | |
2724 | * | |
2725 | * RETURNS: | |
2726 | * Zero on success, non-zero on failure. | |
2727 | */ | |
2728 | static unsigned int atapi_xlat(struct ata_queued_cmd *qc) | |
2729 | { | |
2730 | struct scsi_cmnd *scmd = qc->scsicmd; | |
2731 | struct ata_device *dev = qc->dev; | |
2732 | int nodata = (scmd->sc_data_direction == DMA_NONE); | |
2733 | int using_pio = !nodata && (dev->flags & ATA_DFLAG_PIO); | |
2734 | unsigned int nbytes; | |
2735 | ||
2736 | memset(qc->cdb, 0, dev->cdb_len); | |
2737 | memcpy(qc->cdb, scmd->cmnd, scmd->cmd_len); | |
2738 | ||
2739 | qc->complete_fn = atapi_qc_complete; | |
2740 | ||
2741 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; | |
2742 | if (scmd->sc_data_direction == DMA_TO_DEVICE) { | |
2743 | qc->tf.flags |= ATA_TFLAG_WRITE; | |
2744 | DPRINTK("direction: write\n"); | |
2745 | } | |
2746 | ||
2747 | qc->tf.command = ATA_CMD_PACKET; | |
2748 | ata_qc_set_pc_nbytes(qc); | |
2749 | ||
2750 | /* check whether ATAPI DMA is safe */ | |
2751 | if (!nodata && !using_pio && atapi_check_dma(qc)) | |
2752 | using_pio = 1; | |
2753 | ||
2754 | /* Some controller variants snoop this value for Packet | |
2755 | * transfers to do state machine and FIFO management. Thus we | |
2756 | * want to set it properly, and for DMA where it is | |
2757 | * effectively meaningless. | |
2758 | */ | |
2759 | nbytes = min(ata_qc_raw_nbytes(qc), (unsigned int)63 * 1024); | |
2760 | ||
2761 | /* Most ATAPI devices which honor transfer chunk size don't | |
2762 | * behave according to the spec when odd chunk size which | |
2763 | * matches the transfer length is specified. If the number of | |
2764 | * bytes to transfer is 2n+1. According to the spec, what | |
2765 | * should happen is to indicate that 2n+1 is going to be | |
2766 | * transferred and transfer 2n+2 bytes where the last byte is | |
2767 | * padding. | |
2768 | * | |
2769 | * In practice, this doesn't happen. ATAPI devices first | |
2770 | * indicate and transfer 2n bytes and then indicate and | |
2771 | * transfer 2 bytes where the last byte is padding. | |
2772 | * | |
2773 | * This inconsistency confuses several controllers which | |
2774 | * perform PIO using DMA such as Intel AHCIs and sil3124/32. | |
2775 | * These controllers use actual number of transferred bytes to | |
2776 | * update DMA poitner and transfer of 4n+2 bytes make those | |
2777 | * controller push DMA pointer by 4n+4 bytes because SATA data | |
2778 | * FISes are aligned to 4 bytes. This causes data corruption | |
2779 | * and buffer overrun. | |
2780 | * | |
2781 | * Always setting nbytes to even number solves this problem | |
2782 | * because then ATAPI devices don't have to split data at 2n | |
2783 | * boundaries. | |
2784 | */ | |
2785 | if (nbytes & 0x1) | |
2786 | nbytes++; | |
2787 | ||
2788 | qc->tf.lbam = (nbytes & 0xFF); | |
2789 | qc->tf.lbah = (nbytes >> 8); | |
2790 | ||
2791 | if (nodata) | |
2792 | qc->tf.protocol = ATAPI_PROT_NODATA; | |
2793 | else if (using_pio) | |
2794 | qc->tf.protocol = ATAPI_PROT_PIO; | |
2795 | else { | |
2796 | /* DMA data xfer */ | |
2797 | qc->tf.protocol = ATAPI_PROT_DMA; | |
2798 | qc->tf.feature |= ATAPI_PKT_DMA; | |
2799 | ||
2800 | if ((dev->flags & ATA_DFLAG_DMADIR) && | |
2801 | (scmd->sc_data_direction != DMA_TO_DEVICE)) | |
2802 | /* some SATA bridges need us to indicate data xfer direction */ | |
2803 | qc->tf.feature |= ATAPI_DMADIR; | |
2804 | } | |
2805 | ||
2806 | ||
2807 | /* FIXME: We need to translate 0x05 READ_BLOCK_LIMITS to a MODE_SENSE | |
2808 | as ATAPI tape drives don't get this right otherwise */ | |
2809 | return 0; | |
2810 | } | |
2811 | ||
2812 | static struct ata_device *ata_find_dev(struct ata_port *ap, int devno) | |
2813 | { | |
2814 | if (!sata_pmp_attached(ap)) { | |
2815 | if (likely(devno < ata_link_max_devices(&ap->link))) | |
2816 | return &ap->link.device[devno]; | |
2817 | } else { | |
2818 | if (likely(devno < ap->nr_pmp_links)) | |
2819 | return &ap->pmp_link[devno].device[0]; | |
2820 | } | |
2821 | ||
2822 | return NULL; | |
2823 | } | |
2824 | ||
2825 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, | |
2826 | const struct scsi_device *scsidev) | |
2827 | { | |
2828 | int devno; | |
2829 | ||
2830 | /* skip commands not addressed to targets we simulate */ | |
2831 | if (!sata_pmp_attached(ap)) { | |
2832 | if (unlikely(scsidev->channel || scsidev->lun)) | |
2833 | return NULL; | |
2834 | devno = scsidev->id; | |
2835 | } else { | |
2836 | if (unlikely(scsidev->id || scsidev->lun)) | |
2837 | return NULL; | |
2838 | devno = scsidev->channel; | |
2839 | } | |
2840 | ||
2841 | return ata_find_dev(ap, devno); | |
2842 | } | |
2843 | ||
2844 | /** | |
2845 | * ata_scsi_find_dev - lookup ata_device from scsi_cmnd | |
2846 | * @ap: ATA port to which the device is attached | |
2847 | * @scsidev: SCSI device from which we derive the ATA device | |
2848 | * | |
2849 | * Given various information provided in struct scsi_cmnd, | |
2850 | * map that onto an ATA bus, and using that mapping | |
2851 | * determine which ata_device is associated with the | |
2852 | * SCSI command to be sent. | |
2853 | * | |
2854 | * LOCKING: | |
2855 | * spin_lock_irqsave(host lock) | |
2856 | * | |
2857 | * RETURNS: | |
2858 | * Associated ATA device, or %NULL if not found. | |
2859 | */ | |
2860 | static struct ata_device * | |
2861 | ata_scsi_find_dev(struct ata_port *ap, const struct scsi_device *scsidev) | |
2862 | { | |
2863 | struct ata_device *dev = __ata_scsi_find_dev(ap, scsidev); | |
2864 | ||
2865 | if (unlikely(!dev || !ata_dev_enabled(dev))) | |
2866 | return NULL; | |
2867 | ||
2868 | return dev; | |
2869 | } | |
2870 | ||
2871 | /* | |
2872 | * ata_scsi_map_proto - Map pass-thru protocol value to taskfile value. | |
2873 | * @byte1: Byte 1 from pass-thru CDB. | |
2874 | * | |
2875 | * RETURNS: | |
2876 | * ATA_PROT_UNKNOWN if mapping failed/unimplemented, protocol otherwise. | |
2877 | */ | |
2878 | static u8 | |
2879 | ata_scsi_map_proto(u8 byte1) | |
2880 | { | |
2881 | switch((byte1 & 0x1e) >> 1) { | |
2882 | case 3: /* Non-data */ | |
2883 | return ATA_PROT_NODATA; | |
2884 | ||
2885 | case 6: /* DMA */ | |
2886 | case 10: /* UDMA Data-in */ | |
2887 | case 11: /* UDMA Data-Out */ | |
2888 | return ATA_PROT_DMA; | |
2889 | ||
2890 | case 4: /* PIO Data-in */ | |
2891 | case 5: /* PIO Data-out */ | |
2892 | return ATA_PROT_PIO; | |
2893 | ||
2894 | case 0: /* Hard Reset */ | |
2895 | case 1: /* SRST */ | |
2896 | case 8: /* Device Diagnostic */ | |
2897 | case 9: /* Device Reset */ | |
2898 | case 7: /* DMA Queued */ | |
2899 | case 12: /* FPDMA */ | |
2900 | case 15: /* Return Response Info */ | |
2901 | default: /* Reserved */ | |
2902 | break; | |
2903 | } | |
2904 | ||
2905 | return ATA_PROT_UNKNOWN; | |
2906 | } | |
2907 | ||
2908 | /** | |
2909 | * ata_scsi_pass_thru - convert ATA pass-thru CDB to taskfile | |
2910 | * @qc: command structure to be initialized | |
2911 | * | |
2912 | * Handles either 12 or 16-byte versions of the CDB. | |
2913 | * | |
2914 | * RETURNS: | |
2915 | * Zero on success, non-zero on failure. | |
2916 | */ | |
2917 | static unsigned int ata_scsi_pass_thru(struct ata_queued_cmd *qc) | |
2918 | { | |
2919 | struct ata_taskfile *tf = &(qc->tf); | |
2920 | struct scsi_cmnd *scmd = qc->scsicmd; | |
2921 | struct ata_device *dev = qc->dev; | |
2922 | const u8 *cdb = scmd->cmnd; | |
2923 | ||
2924 | if ((tf->protocol = ata_scsi_map_proto(cdb[1])) == ATA_PROT_UNKNOWN) | |
2925 | goto invalid_fld; | |
2926 | ||
2927 | /* | |
2928 | * 12 and 16 byte CDBs use different offsets to | |
2929 | * provide the various register values. | |
2930 | */ | |
2931 | if (cdb[0] == ATA_16) { | |
2932 | /* | |
2933 | * 16-byte CDB - may contain extended commands. | |
2934 | * | |
2935 | * If that is the case, copy the upper byte register values. | |
2936 | */ | |
2937 | if (cdb[1] & 0x01) { | |
2938 | tf->hob_feature = cdb[3]; | |
2939 | tf->hob_nsect = cdb[5]; | |
2940 | tf->hob_lbal = cdb[7]; | |
2941 | tf->hob_lbam = cdb[9]; | |
2942 | tf->hob_lbah = cdb[11]; | |
2943 | tf->flags |= ATA_TFLAG_LBA48; | |
2944 | } else | |
2945 | tf->flags &= ~ATA_TFLAG_LBA48; | |
2946 | ||
2947 | /* | |
2948 | * Always copy low byte, device and command registers. | |
2949 | */ | |
2950 | tf->feature = cdb[4]; | |
2951 | tf->nsect = cdb[6]; | |
2952 | tf->lbal = cdb[8]; | |
2953 | tf->lbam = cdb[10]; | |
2954 | tf->lbah = cdb[12]; | |
2955 | tf->device = cdb[13]; | |
2956 | tf->command = cdb[14]; | |
2957 | } else { | |
2958 | /* | |
2959 | * 12-byte CDB - incapable of extended commands. | |
2960 | */ | |
2961 | tf->flags &= ~ATA_TFLAG_LBA48; | |
2962 | ||
2963 | tf->feature = cdb[3]; | |
2964 | tf->nsect = cdb[4]; | |
2965 | tf->lbal = cdb[5]; | |
2966 | tf->lbam = cdb[6]; | |
2967 | tf->lbah = cdb[7]; | |
2968 | tf->device = cdb[8]; | |
2969 | tf->command = cdb[9]; | |
2970 | } | |
2971 | ||
2972 | /* enforce correct master/slave bit */ | |
2973 | tf->device = dev->devno ? | |
2974 | tf->device | ATA_DEV1 : tf->device & ~ATA_DEV1; | |
2975 | ||
2976 | switch (tf->command) { | |
2977 | /* READ/WRITE LONG use a non-standard sect_size */ | |
2978 | case ATA_CMD_READ_LONG: | |
2979 | case ATA_CMD_READ_LONG_ONCE: | |
2980 | case ATA_CMD_WRITE_LONG: | |
2981 | case ATA_CMD_WRITE_LONG_ONCE: | |
2982 | if (tf->protocol != ATA_PROT_PIO || tf->nsect != 1) | |
2983 | goto invalid_fld; | |
2984 | qc->sect_size = scsi_bufflen(scmd); | |
2985 | break; | |
2986 | ||
2987 | /* commands using reported Logical Block size (e.g. 512 or 4K) */ | |
2988 | case ATA_CMD_CFA_WRITE_NE: | |
2989 | case ATA_CMD_CFA_TRANS_SECT: | |
2990 | case ATA_CMD_CFA_WRITE_MULT_NE: | |
2991 | /* XXX: case ATA_CMD_CFA_WRITE_SECTORS_WITHOUT_ERASE: */ | |
2992 | case ATA_CMD_READ: | |
2993 | case ATA_CMD_READ_EXT: | |
2994 | case ATA_CMD_READ_QUEUED: | |
2995 | /* XXX: case ATA_CMD_READ_QUEUED_EXT: */ | |
2996 | case ATA_CMD_FPDMA_READ: | |
2997 | case ATA_CMD_READ_MULTI: | |
2998 | case ATA_CMD_READ_MULTI_EXT: | |
2999 | case ATA_CMD_PIO_READ: | |
3000 | case ATA_CMD_PIO_READ_EXT: | |
3001 | case ATA_CMD_READ_STREAM_DMA_EXT: | |
3002 | case ATA_CMD_READ_STREAM_EXT: | |
3003 | case ATA_CMD_VERIFY: | |
3004 | case ATA_CMD_VERIFY_EXT: | |
3005 | case ATA_CMD_WRITE: | |
3006 | case ATA_CMD_WRITE_EXT: | |
3007 | case ATA_CMD_WRITE_FUA_EXT: | |
3008 | case ATA_CMD_WRITE_QUEUED: | |
3009 | case ATA_CMD_WRITE_QUEUED_FUA_EXT: | |
3010 | case ATA_CMD_FPDMA_WRITE: | |
3011 | case ATA_CMD_WRITE_MULTI: | |
3012 | case ATA_CMD_WRITE_MULTI_EXT: | |
3013 | case ATA_CMD_WRITE_MULTI_FUA_EXT: | |
3014 | case ATA_CMD_PIO_WRITE: | |
3015 | case ATA_CMD_PIO_WRITE_EXT: | |
3016 | case ATA_CMD_WRITE_STREAM_DMA_EXT: | |
3017 | case ATA_CMD_WRITE_STREAM_EXT: | |
3018 | qc->sect_size = scmd->device->sector_size; | |
3019 | break; | |
3020 | ||
3021 | /* Everything else uses 512 byte "sectors" */ | |
3022 | default: | |
3023 | qc->sect_size = ATA_SECT_SIZE; | |
3024 | } | |
3025 | ||
3026 | /* | |
3027 | * Set flags so that all registers will be written, pass on | |
3028 | * write indication (used for PIO/DMA setup), result TF is | |
3029 | * copied back and we don't whine too much about its failure. | |
3030 | */ | |
3031 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; | |
3032 | if (scmd->sc_data_direction == DMA_TO_DEVICE) | |
3033 | tf->flags |= ATA_TFLAG_WRITE; | |
3034 | ||
3035 | qc->flags |= ATA_QCFLAG_RESULT_TF | ATA_QCFLAG_QUIET; | |
3036 | ||
3037 | /* | |
3038 | * Set transfer length. | |
3039 | * | |
3040 | * TODO: find out if we need to do more here to | |
3041 | * cover scatter/gather case. | |
3042 | */ | |
3043 | ata_qc_set_pc_nbytes(qc); | |
3044 | ||
3045 | /* We may not issue DMA commands if no DMA mode is set */ | |
3046 | if (tf->protocol == ATA_PROT_DMA && dev->dma_mode == 0) | |
3047 | goto invalid_fld; | |
3048 | ||
3049 | /* sanity check for pio multi commands */ | |
3050 | if ((cdb[1] & 0xe0) && !is_multi_taskfile(tf)) | |
3051 | goto invalid_fld; | |
3052 | ||
3053 | if (is_multi_taskfile(tf)) { | |
3054 | unsigned int multi_count = 1 << (cdb[1] >> 5); | |
3055 | ||
3056 | /* compare the passed through multi_count | |
3057 | * with the cached multi_count of libata | |
3058 | */ | |
3059 | if (multi_count != dev->multi_count) | |
3060 | ata_dev_warn(dev, "invalid multi_count %u ignored\n", | |
3061 | multi_count); | |
3062 | } | |
3063 | ||
3064 | /* | |
3065 | * Filter SET_FEATURES - XFER MODE command -- otherwise, | |
3066 | * SET_FEATURES - XFER MODE must be preceded/succeeded | |
3067 | * by an update to hardware-specific registers for each | |
3068 | * controller (i.e. the reason for ->set_piomode(), | |
3069 | * ->set_dmamode(), and ->post_set_mode() hooks). | |
3070 | */ | |
3071 | if (tf->command == ATA_CMD_SET_FEATURES && | |
3072 | tf->feature == SETFEATURES_XFER) | |
3073 | goto invalid_fld; | |
3074 | ||
3075 | /* | |
3076 | * Filter TPM commands by default. These provide an | |
3077 | * essentially uncontrolled encrypted "back door" between | |
3078 | * applications and the disk. Set libata.allow_tpm=1 if you | |
3079 | * have a real reason for wanting to use them. This ensures | |
3080 | * that installed software cannot easily mess stuff up without | |
3081 | * user intent. DVR type users will probably ship with this enabled | |
3082 | * for movie content management. | |
3083 | * | |
3084 | * Note that for ATA8 we can issue a DCS change and DCS freeze lock | |
3085 | * for this and should do in future but that it is not sufficient as | |
3086 | * DCS is an optional feature set. Thus we also do the software filter | |
3087 | * so that we comply with the TC consortium stated goal that the user | |
3088 | * can turn off TC features of their system. | |
3089 | */ | |
3090 | if (tf->command >= 0x5C && tf->command <= 0x5F && !libata_allow_tpm) | |
3091 | goto invalid_fld; | |
3092 | ||
3093 | return 0; | |
3094 | ||
3095 | invalid_fld: | |
3096 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); | |
3097 | /* "Invalid field in cdb" */ | |
3098 | return 1; | |
3099 | } | |
3100 | ||
3101 | static unsigned int ata_scsi_write_same_xlat(struct ata_queued_cmd *qc) | |
3102 | { | |
3103 | struct ata_taskfile *tf = &qc->tf; | |
3104 | struct scsi_cmnd *scmd = qc->scsicmd; | |
3105 | struct ata_device *dev = qc->dev; | |
3106 | const u8 *cdb = scmd->cmnd; | |
3107 | u64 block; | |
3108 | u32 n_block; | |
3109 | u32 size; | |
3110 | void *buf; | |
3111 | ||
3112 | /* we may not issue DMA commands if no DMA mode is set */ | |
3113 | if (unlikely(!dev->dma_mode)) | |
3114 | goto invalid_fld; | |
3115 | ||
3116 | if (unlikely(scmd->cmd_len < 16)) | |
3117 | goto invalid_fld; | |
3118 | scsi_16_lba_len(cdb, &block, &n_block); | |
3119 | ||
3120 | /* for now we only support WRITE SAME with the unmap bit set */ | |
3121 | if (unlikely(!(cdb[1] & 0x8))) | |
3122 | goto invalid_fld; | |
3123 | ||
3124 | /* | |
3125 | * WRITE SAME always has a sector sized buffer as payload, this | |
3126 | * should never be a multiple entry S/G list. | |
3127 | */ | |
3128 | if (!scsi_sg_count(scmd)) | |
3129 | goto invalid_fld; | |
3130 | ||
3131 | buf = page_address(sg_page(scsi_sglist(scmd))); | |
3132 | size = ata_set_lba_range_entries(buf, 512, block, n_block); | |
3133 | ||
3134 | if (ata_ncq_enabled(dev) && ata_fpdma_dsm_supported(dev)) { | |
3135 | /* Newer devices support queued TRIM commands */ | |
3136 | tf->protocol = ATA_PROT_NCQ; | |
3137 | tf->command = ATA_CMD_FPDMA_SEND; | |
3138 | tf->hob_nsect = ATA_SUBCMD_FPDMA_SEND_DSM & 0x1f; | |
3139 | tf->nsect = qc->tag << 3; | |
3140 | tf->hob_feature = (size / 512) >> 8; | |
3141 | tf->feature = size / 512; | |
3142 | ||
3143 | tf->auxiliary = 1; | |
3144 | } else { | |
3145 | tf->protocol = ATA_PROT_DMA; | |
3146 | tf->hob_feature = 0; | |
3147 | tf->feature = ATA_DSM_TRIM; | |
3148 | tf->hob_nsect = (size / 512) >> 8; | |
3149 | tf->nsect = size / 512; | |
3150 | tf->command = ATA_CMD_DSM; | |
3151 | } | |
3152 | ||
3153 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE | ATA_TFLAG_LBA48 | | |
3154 | ATA_TFLAG_WRITE; | |
3155 | ||
3156 | ata_qc_set_pc_nbytes(qc); | |
3157 | ||
3158 | return 0; | |
3159 | ||
3160 | invalid_fld: | |
3161 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); | |
3162 | /* "Invalid field in cdb" */ | |
3163 | return 1; | |
3164 | } | |
3165 | ||
3166 | /** | |
3167 | * ata_mselect_caching - Simulate MODE SELECT for caching info page | |
3168 | * @qc: Storage for translated ATA taskfile | |
3169 | * @buf: input buffer | |
3170 | * @len: number of valid bytes in the input buffer | |
3171 | * | |
3172 | * Prepare a taskfile to modify caching information for the device. | |
3173 | * | |
3174 | * LOCKING: | |
3175 | * None. | |
3176 | */ | |
3177 | static int ata_mselect_caching(struct ata_queued_cmd *qc, | |
3178 | const u8 *buf, int len) | |
3179 | { | |
3180 | struct ata_taskfile *tf = &qc->tf; | |
3181 | struct ata_device *dev = qc->dev; | |
3182 | char mpage[CACHE_MPAGE_LEN]; | |
3183 | u8 wce; | |
3184 | ||
3185 | /* | |
3186 | * The first two bytes of def_cache_mpage are a header, so offsets | |
3187 | * in mpage are off by 2 compared to buf. Same for len. | |
3188 | */ | |
3189 | ||
3190 | if (len != CACHE_MPAGE_LEN - 2) | |
3191 | return -EINVAL; | |
3192 | ||
3193 | wce = buf[0] & (1 << 2); | |
3194 | ||
3195 | /* | |
3196 | * Check that read-only bits are not modified. | |
3197 | */ | |
3198 | ata_msense_caching(dev->id, mpage, false); | |
3199 | mpage[2] &= ~(1 << 2); | |
3200 | mpage[2] |= wce; | |
3201 | if (memcmp(mpage + 2, buf, CACHE_MPAGE_LEN - 2) != 0) | |
3202 | return -EINVAL; | |
3203 | ||
3204 | tf->flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; | |
3205 | tf->protocol = ATA_PROT_NODATA; | |
3206 | tf->nsect = 0; | |
3207 | tf->command = ATA_CMD_SET_FEATURES; | |
3208 | tf->feature = wce ? SETFEATURES_WC_ON : SETFEATURES_WC_OFF; | |
3209 | return 0; | |
3210 | } | |
3211 | ||
3212 | /** | |
3213 | * ata_scsiop_mode_select - Simulate MODE SELECT 6, 10 commands | |
3214 | * @qc: Storage for translated ATA taskfile | |
3215 | * | |
3216 | * Converts a MODE SELECT command to an ATA SET FEATURES taskfile. | |
3217 | * Assume this is invoked for direct access devices (e.g. disks) only. | |
3218 | * There should be no block descriptor for other device types. | |
3219 | * | |
3220 | * LOCKING: | |
3221 | * spin_lock_irqsave(host lock) | |
3222 | */ | |
3223 | static unsigned int ata_scsi_mode_select_xlat(struct ata_queued_cmd *qc) | |
3224 | { | |
3225 | struct scsi_cmnd *scmd = qc->scsicmd; | |
3226 | const u8 *cdb = scmd->cmnd; | |
3227 | const u8 *p; | |
3228 | u8 pg, spg; | |
3229 | unsigned six_byte, pg_len, hdr_len, bd_len; | |
3230 | int len; | |
3231 | ||
3232 | VPRINTK("ENTER\n"); | |
3233 | ||
3234 | six_byte = (cdb[0] == MODE_SELECT); | |
3235 | if (six_byte) { | |
3236 | if (scmd->cmd_len < 5) | |
3237 | goto invalid_fld; | |
3238 | ||
3239 | len = cdb[4]; | |
3240 | hdr_len = 4; | |
3241 | } else { | |
3242 | if (scmd->cmd_len < 9) | |
3243 | goto invalid_fld; | |
3244 | ||
3245 | len = (cdb[7] << 8) + cdb[8]; | |
3246 | hdr_len = 8; | |
3247 | } | |
3248 | ||
3249 | /* We only support PF=1, SP=0. */ | |
3250 | if ((cdb[1] & 0x11) != 0x10) | |
3251 | goto invalid_fld; | |
3252 | ||
3253 | /* Test early for possible overrun. */ | |
3254 | if (!scsi_sg_count(scmd) || scsi_sglist(scmd)->length < len) | |
3255 | goto invalid_param_len; | |
3256 | ||
3257 | p = page_address(sg_page(scsi_sglist(scmd))); | |
3258 | ||
3259 | /* Move past header and block descriptors. */ | |
3260 | if (len < hdr_len) | |
3261 | goto invalid_param_len; | |
3262 | ||
3263 | if (six_byte) | |
3264 | bd_len = p[3]; | |
3265 | else | |
3266 | bd_len = (p[6] << 8) + p[7]; | |
3267 | ||
3268 | len -= hdr_len; | |
3269 | p += hdr_len; | |
3270 | if (len < bd_len) | |
3271 | goto invalid_param_len; | |
3272 | if (bd_len != 0 && bd_len != 8) | |
3273 | goto invalid_param; | |
3274 | ||
3275 | len -= bd_len; | |
3276 | p += bd_len; | |
3277 | if (len == 0) | |
3278 | goto skip; | |
3279 | ||
3280 | /* Parse both possible formats for the mode page headers. */ | |
3281 | pg = p[0] & 0x3f; | |
3282 | if (p[0] & 0x40) { | |
3283 | if (len < 4) | |
3284 | goto invalid_param_len; | |
3285 | ||
3286 | spg = p[1]; | |
3287 | pg_len = (p[2] << 8) | p[3]; | |
3288 | p += 4; | |
3289 | len -= 4; | |
3290 | } else { | |
3291 | if (len < 2) | |
3292 | goto invalid_param_len; | |
3293 | ||
3294 | spg = 0; | |
3295 | pg_len = p[1]; | |
3296 | p += 2; | |
3297 | len -= 2; | |
3298 | } | |
3299 | ||
3300 | /* | |
3301 | * No mode subpages supported (yet) but asking for _all_ | |
3302 | * subpages may be valid | |
3303 | */ | |
3304 | if (spg && (spg != ALL_SUB_MPAGES)) | |
3305 | goto invalid_param; | |
3306 | if (pg_len > len) | |
3307 | goto invalid_param_len; | |
3308 | ||
3309 | switch (pg) { | |
3310 | case CACHE_MPAGE: | |
3311 | if (ata_mselect_caching(qc, p, pg_len) < 0) | |
3312 | goto invalid_param; | |
3313 | break; | |
3314 | ||
3315 | default: /* invalid page code */ | |
3316 | goto invalid_param; | |
3317 | } | |
3318 | ||
3319 | /* | |
3320 | * Only one page has changeable data, so we only support setting one | |
3321 | * page at a time. | |
3322 | */ | |
3323 | if (len > pg_len) | |
3324 | goto invalid_param; | |
3325 | ||
3326 | return 0; | |
3327 | ||
3328 | invalid_fld: | |
3329 | /* "Invalid field in CDB" */ | |
3330 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); | |
3331 | return 1; | |
3332 | ||
3333 | invalid_param: | |
3334 | /* "Invalid field in parameter list" */ | |
3335 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x26, 0x0); | |
3336 | return 1; | |
3337 | ||
3338 | invalid_param_len: | |
3339 | /* "Parameter list length error" */ | |
3340 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x1a, 0x0); | |
3341 | return 1; | |
3342 | ||
3343 | skip: | |
3344 | scmd->result = SAM_STAT_GOOD; | |
3345 | return 1; | |
3346 | } | |
3347 | ||
3348 | /** | |
3349 | * ata_get_xlat_func - check if SCSI to ATA translation is possible | |
3350 | * @dev: ATA device | |
3351 | * @cmd: SCSI command opcode to consider | |
3352 | * | |
3353 | * Look up the SCSI command given, and determine whether the | |
3354 | * SCSI command is to be translated or simulated. | |
3355 | * | |
3356 | * RETURNS: | |
3357 | * Pointer to translation function if possible, %NULL if not. | |
3358 | */ | |
3359 | ||
3360 | static inline ata_xlat_func_t ata_get_xlat_func(struct ata_device *dev, u8 cmd) | |
3361 | { | |
3362 | switch (cmd) { | |
3363 | case READ_6: | |
3364 | case READ_10: | |
3365 | case READ_16: | |
3366 | ||
3367 | case WRITE_6: | |
3368 | case WRITE_10: | |
3369 | case WRITE_16: | |
3370 | return ata_scsi_rw_xlat; | |
3371 | ||
3372 | case WRITE_SAME_16: | |
3373 | return ata_scsi_write_same_xlat; | |
3374 | ||
3375 | case SYNCHRONIZE_CACHE: | |
3376 | if (ata_try_flush_cache(dev)) | |
3377 | return ata_scsi_flush_xlat; | |
3378 | break; | |
3379 | ||
3380 | case VERIFY: | |
3381 | case VERIFY_16: | |
3382 | return ata_scsi_verify_xlat; | |
3383 | ||
3384 | case ATA_12: | |
3385 | case ATA_16: | |
3386 | return ata_scsi_pass_thru; | |
3387 | ||
3388 | case MODE_SELECT: | |
3389 | case MODE_SELECT_10: | |
3390 | return ata_scsi_mode_select_xlat; | |
3391 | break; | |
3392 | ||
3393 | case START_STOP: | |
3394 | return ata_scsi_start_stop_xlat; | |
3395 | } | |
3396 | ||
3397 | return NULL; | |
3398 | } | |
3399 | ||
3400 | /** | |
3401 | * ata_scsi_dump_cdb - dump SCSI command contents to dmesg | |
3402 | * @ap: ATA port to which the command was being sent | |
3403 | * @cmd: SCSI command to dump | |
3404 | * | |
3405 | * Prints the contents of a SCSI command via printk(). | |
3406 | */ | |
3407 | ||
3408 | static inline void ata_scsi_dump_cdb(struct ata_port *ap, | |
3409 | struct scsi_cmnd *cmd) | |
3410 | { | |
3411 | #ifdef ATA_DEBUG | |
3412 | struct scsi_device *scsidev = cmd->device; | |
3413 | u8 *scsicmd = cmd->cmnd; | |
3414 | ||
3415 | DPRINTK("CDB (%u:%d,%d,%d) %02x %02x %02x %02x %02x %02x %02x %02x %02x\n", | |
3416 | ap->print_id, | |
3417 | scsidev->channel, scsidev->id, scsidev->lun, | |
3418 | scsicmd[0], scsicmd[1], scsicmd[2], scsicmd[3], | |
3419 | scsicmd[4], scsicmd[5], scsicmd[6], scsicmd[7], | |
3420 | scsicmd[8]); | |
3421 | #endif | |
3422 | } | |
3423 | ||
3424 | static inline int __ata_scsi_queuecmd(struct scsi_cmnd *scmd, | |
3425 | struct ata_device *dev) | |
3426 | { | |
3427 | u8 scsi_op = scmd->cmnd[0]; | |
3428 | ata_xlat_func_t xlat_func; | |
3429 | int rc = 0; | |
3430 | ||
3431 | if (dev->class == ATA_DEV_ATA || dev->class == ATA_DEV_ZAC) { | |
3432 | if (unlikely(!scmd->cmd_len || scmd->cmd_len > dev->cdb_len)) | |
3433 | goto bad_cdb_len; | |
3434 | ||
3435 | xlat_func = ata_get_xlat_func(dev, scsi_op); | |
3436 | } else { | |
3437 | if (unlikely(!scmd->cmd_len)) | |
3438 | goto bad_cdb_len; | |
3439 | ||
3440 | xlat_func = NULL; | |
3441 | if (likely((scsi_op != ATA_16) || !atapi_passthru16)) { | |
3442 | /* relay SCSI command to ATAPI device */ | |
3443 | int len = COMMAND_SIZE(scsi_op); | |
3444 | if (unlikely(len > scmd->cmd_len || len > dev->cdb_len)) | |
3445 | goto bad_cdb_len; | |
3446 | ||
3447 | xlat_func = atapi_xlat; | |
3448 | } else { | |
3449 | /* ATA_16 passthru, treat as an ATA command */ | |
3450 | if (unlikely(scmd->cmd_len > 16)) | |
3451 | goto bad_cdb_len; | |
3452 | ||
3453 | xlat_func = ata_get_xlat_func(dev, scsi_op); | |
3454 | } | |
3455 | } | |
3456 | ||
3457 | if (xlat_func) | |
3458 | rc = ata_scsi_translate(dev, scmd, xlat_func); | |
3459 | else | |
3460 | ata_scsi_simulate(dev, scmd); | |
3461 | ||
3462 | return rc; | |
3463 | ||
3464 | bad_cdb_len: | |
3465 | DPRINTK("bad CDB len=%u, scsi_op=0x%02x, max=%u\n", | |
3466 | scmd->cmd_len, scsi_op, dev->cdb_len); | |
3467 | scmd->result = DID_ERROR << 16; | |
3468 | scmd->scsi_done(scmd); | |
3469 | return 0; | |
3470 | } | |
3471 | ||
3472 | /** | |
3473 | * ata_scsi_queuecmd - Issue SCSI cdb to libata-managed device | |
3474 | * @shost: SCSI host of command to be sent | |
3475 | * @cmd: SCSI command to be sent | |
3476 | * | |
3477 | * In some cases, this function translates SCSI commands into | |
3478 | * ATA taskfiles, and queues the taskfiles to be sent to | |
3479 | * hardware. In other cases, this function simulates a | |
3480 | * SCSI device by evaluating and responding to certain | |
3481 | * SCSI commands. This creates the overall effect of | |
3482 | * ATA and ATAPI devices appearing as SCSI devices. | |
3483 | * | |
3484 | * LOCKING: | |
3485 | * ATA host lock | |
3486 | * | |
3487 | * RETURNS: | |
3488 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, | |
3489 | * 0 otherwise. | |
3490 | */ | |
3491 | int ata_scsi_queuecmd(struct Scsi_Host *shost, struct scsi_cmnd *cmd) | |
3492 | { | |
3493 | struct ata_port *ap; | |
3494 | struct ata_device *dev; | |
3495 | struct scsi_device *scsidev = cmd->device; | |
3496 | int rc = 0; | |
3497 | unsigned long irq_flags; | |
3498 | ||
3499 | ap = ata_shost_to_port(shost); | |
3500 | ||
3501 | spin_lock_irqsave(ap->lock, irq_flags); | |
3502 | ||
3503 | ata_scsi_dump_cdb(ap, cmd); | |
3504 | ||
3505 | dev = ata_scsi_find_dev(ap, scsidev); | |
3506 | if (likely(dev)) | |
3507 | rc = __ata_scsi_queuecmd(cmd, dev); | |
3508 | else { | |
3509 | cmd->result = (DID_BAD_TARGET << 16); | |
3510 | cmd->scsi_done(cmd); | |
3511 | } | |
3512 | ||
3513 | spin_unlock_irqrestore(ap->lock, irq_flags); | |
3514 | ||
3515 | return rc; | |
3516 | } | |
3517 | ||
3518 | /** | |
3519 | * ata_scsi_simulate - simulate SCSI command on ATA device | |
3520 | * @dev: the target device | |
3521 | * @cmd: SCSI command being sent to device. | |
3522 | * | |
3523 | * Interprets and directly executes a select list of SCSI commands | |
3524 | * that can be handled internally. | |
3525 | * | |
3526 | * LOCKING: | |
3527 | * spin_lock_irqsave(host lock) | |
3528 | */ | |
3529 | ||
3530 | void ata_scsi_simulate(struct ata_device *dev, struct scsi_cmnd *cmd) | |
3531 | { | |
3532 | struct ata_scsi_args args; | |
3533 | const u8 *scsicmd = cmd->cmnd; | |
3534 | u8 tmp8; | |
3535 | ||
3536 | args.dev = dev; | |
3537 | args.id = dev->id; | |
3538 | args.cmd = cmd; | |
3539 | args.done = cmd->scsi_done; | |
3540 | ||
3541 | switch(scsicmd[0]) { | |
3542 | /* TODO: worth improving? */ | |
3543 | case FORMAT_UNIT: | |
3544 | ata_scsi_invalid_field(cmd); | |
3545 | break; | |
3546 | ||
3547 | case INQUIRY: | |
3548 | if (scsicmd[1] & 2) /* is CmdDt set? */ | |
3549 | ata_scsi_invalid_field(cmd); | |
3550 | else if ((scsicmd[1] & 1) == 0) /* is EVPD clear? */ | |
3551 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_std); | |
3552 | else switch (scsicmd[2]) { | |
3553 | case 0x00: | |
3554 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_00); | |
3555 | break; | |
3556 | case 0x80: | |
3557 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_80); | |
3558 | break; | |
3559 | case 0x83: | |
3560 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_83); | |
3561 | break; | |
3562 | case 0x89: | |
3563 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_89); | |
3564 | break; | |
3565 | case 0xb0: | |
3566 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b0); | |
3567 | break; | |
3568 | case 0xb1: | |
3569 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b1); | |
3570 | break; | |
3571 | case 0xb2: | |
3572 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b2); | |
3573 | break; | |
3574 | default: | |
3575 | ata_scsi_invalid_field(cmd); | |
3576 | break; | |
3577 | } | |
3578 | break; | |
3579 | ||
3580 | case MODE_SENSE: | |
3581 | case MODE_SENSE_10: | |
3582 | ata_scsi_rbuf_fill(&args, ata_scsiop_mode_sense); | |
3583 | break; | |
3584 | ||
3585 | case READ_CAPACITY: | |
3586 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); | |
3587 | break; | |
3588 | ||
3589 | case SERVICE_ACTION_IN_16: | |
3590 | if ((scsicmd[1] & 0x1f) == SAI_READ_CAPACITY_16) | |
3591 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); | |
3592 | else | |
3593 | ata_scsi_invalid_field(cmd); | |
3594 | break; | |
3595 | ||
3596 | case REPORT_LUNS: | |
3597 | ata_scsi_rbuf_fill(&args, ata_scsiop_report_luns); | |
3598 | break; | |
3599 | ||
3600 | case REQUEST_SENSE: | |
3601 | ata_scsi_set_sense(cmd, 0, 0, 0); | |
3602 | cmd->result = (DRIVER_SENSE << 24); | |
3603 | cmd->scsi_done(cmd); | |
3604 | break; | |
3605 | ||
3606 | /* if we reach this, then writeback caching is disabled, | |
3607 | * turning this into a no-op. | |
3608 | */ | |
3609 | case SYNCHRONIZE_CACHE: | |
3610 | /* fall through */ | |
3611 | ||
3612 | /* no-op's, complete with success */ | |
3613 | case REZERO_UNIT: | |
3614 | case SEEK_6: | |
3615 | case SEEK_10: | |
3616 | case TEST_UNIT_READY: | |
3617 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); | |
3618 | break; | |
3619 | ||
3620 | case SEND_DIAGNOSTIC: | |
3621 | tmp8 = scsicmd[1] & ~(1 << 3); | |
3622 | if ((tmp8 == 0x4) && (!scsicmd[3]) && (!scsicmd[4])) | |
3623 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); | |
3624 | else | |
3625 | ata_scsi_invalid_field(cmd); | |
3626 | break; | |
3627 | ||
3628 | /* all other commands */ | |
3629 | default: | |
3630 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x20, 0x0); | |
3631 | /* "Invalid command operation code" */ | |
3632 | cmd->scsi_done(cmd); | |
3633 | break; | |
3634 | } | |
3635 | } | |
3636 | ||
3637 | int ata_scsi_add_hosts(struct ata_host *host, struct scsi_host_template *sht) | |
3638 | { | |
3639 | int i, rc; | |
3640 | ||
3641 | for (i = 0; i < host->n_ports; i++) { | |
3642 | struct ata_port *ap = host->ports[i]; | |
3643 | struct Scsi_Host *shost; | |
3644 | ||
3645 | rc = -ENOMEM; | |
3646 | shost = scsi_host_alloc(sht, sizeof(struct ata_port *)); | |
3647 | if (!shost) | |
3648 | goto err_alloc; | |
3649 | ||
3650 | shost->eh_noresume = 1; | |
3651 | *(struct ata_port **)&shost->hostdata[0] = ap; | |
3652 | ap->scsi_host = shost; | |
3653 | ||
3654 | shost->transportt = ata_scsi_transport_template; | |
3655 | shost->unique_id = ap->print_id; | |
3656 | shost->max_id = 16; | |
3657 | shost->max_lun = 1; | |
3658 | shost->max_channel = 1; | |
3659 | shost->max_cmd_len = 16; | |
3660 | shost->no_write_same = 1; | |
3661 | ||
3662 | /* Schedule policy is determined by ->qc_defer() | |
3663 | * callback and it needs to see every deferred qc. | |
3664 | * Set host_blocked to 1 to prevent SCSI midlayer from | |
3665 | * automatically deferring requests. | |
3666 | */ | |
3667 | shost->max_host_blocked = 1; | |
3668 | ||
3669 | rc = scsi_add_host_with_dma(ap->scsi_host, | |
3670 | &ap->tdev, ap->host->dev); | |
3671 | if (rc) | |
3672 | goto err_add; | |
3673 | } | |
3674 | ||
3675 | return 0; | |
3676 | ||
3677 | err_add: | |
3678 | scsi_host_put(host->ports[i]->scsi_host); | |
3679 | err_alloc: | |
3680 | while (--i >= 0) { | |
3681 | struct Scsi_Host *shost = host->ports[i]->scsi_host; | |
3682 | ||
3683 | scsi_remove_host(shost); | |
3684 | scsi_host_put(shost); | |
3685 | } | |
3686 | return rc; | |
3687 | } | |
3688 | ||
3689 | void ata_scsi_scan_host(struct ata_port *ap, int sync) | |
3690 | { | |
3691 | int tries = 5; | |
3692 | struct ata_device *last_failed_dev = NULL; | |
3693 | struct ata_link *link; | |
3694 | struct ata_device *dev; | |
3695 | ||
3696 | repeat: | |
3697 | ata_for_each_link(link, ap, EDGE) { | |
3698 | ata_for_each_dev(dev, link, ENABLED) { | |
3699 | struct scsi_device *sdev; | |
3700 | int channel = 0, id = 0; | |
3701 | ||
3702 | if (dev->sdev) | |
3703 | continue; | |
3704 | ||
3705 | if (ata_is_host_link(link)) | |
3706 | id = dev->devno; | |
3707 | else | |
3708 | channel = link->pmp; | |
3709 | ||
3710 | sdev = __scsi_add_device(ap->scsi_host, channel, id, 0, | |
3711 | NULL); | |
3712 | if (!IS_ERR(sdev)) { | |
3713 | dev->sdev = sdev; | |
3714 | scsi_device_put(sdev); | |
3715 | } else { | |
3716 | dev->sdev = NULL; | |
3717 | } | |
3718 | } | |
3719 | } | |
3720 | ||
3721 | /* If we scanned while EH was in progress or allocation | |
3722 | * failure occurred, scan would have failed silently. Check | |
3723 | * whether all devices are attached. | |
3724 | */ | |
3725 | ata_for_each_link(link, ap, EDGE) { | |
3726 | ata_for_each_dev(dev, link, ENABLED) { | |
3727 | if (!dev->sdev) | |
3728 | goto exit_loop; | |
3729 | } | |
3730 | } | |
3731 | exit_loop: | |
3732 | if (!link) | |
3733 | return; | |
3734 | ||
3735 | /* we're missing some SCSI devices */ | |
3736 | if (sync) { | |
3737 | /* If caller requested synchrnous scan && we've made | |
3738 | * any progress, sleep briefly and repeat. | |
3739 | */ | |
3740 | if (dev != last_failed_dev) { | |
3741 | msleep(100); | |
3742 | last_failed_dev = dev; | |
3743 | goto repeat; | |
3744 | } | |
3745 | ||
3746 | /* We might be failing to detect boot device, give it | |
3747 | * a few more chances. | |
3748 | */ | |
3749 | if (--tries) { | |
3750 | msleep(100); | |
3751 | goto repeat; | |
3752 | } | |
3753 | ||
3754 | ata_port_err(ap, | |
3755 | "WARNING: synchronous SCSI scan failed without making any progress, switching to async\n"); | |
3756 | } | |
3757 | ||
3758 | queue_delayed_work(system_long_wq, &ap->hotplug_task, | |
3759 | round_jiffies_relative(HZ)); | |
3760 | } | |
3761 | ||
3762 | /** | |
3763 | * ata_scsi_offline_dev - offline attached SCSI device | |
3764 | * @dev: ATA device to offline attached SCSI device for | |
3765 | * | |
3766 | * This function is called from ata_eh_hotplug() and responsible | |
3767 | * for taking the SCSI device attached to @dev offline. This | |
3768 | * function is called with host lock which protects dev->sdev | |
3769 | * against clearing. | |
3770 | * | |
3771 | * LOCKING: | |
3772 | * spin_lock_irqsave(host lock) | |
3773 | * | |
3774 | * RETURNS: | |
3775 | * 1 if attached SCSI device exists, 0 otherwise. | |
3776 | */ | |
3777 | int ata_scsi_offline_dev(struct ata_device *dev) | |
3778 | { | |
3779 | if (dev->sdev) { | |
3780 | scsi_device_set_state(dev->sdev, SDEV_OFFLINE); | |
3781 | return 1; | |
3782 | } | |
3783 | return 0; | |
3784 | } | |
3785 | ||
3786 | /** | |
3787 | * ata_scsi_remove_dev - remove attached SCSI device | |
3788 | * @dev: ATA device to remove attached SCSI device for | |
3789 | * | |
3790 | * This function is called from ata_eh_scsi_hotplug() and | |
3791 | * responsible for removing the SCSI device attached to @dev. | |
3792 | * | |
3793 | * LOCKING: | |
3794 | * Kernel thread context (may sleep). | |
3795 | */ | |
3796 | static void ata_scsi_remove_dev(struct ata_device *dev) | |
3797 | { | |
3798 | struct ata_port *ap = dev->link->ap; | |
3799 | struct scsi_device *sdev; | |
3800 | unsigned long flags; | |
3801 | ||
3802 | /* Alas, we need to grab scan_mutex to ensure SCSI device | |
3803 | * state doesn't change underneath us and thus | |
3804 | * scsi_device_get() always succeeds. The mutex locking can | |
3805 | * be removed if there is __scsi_device_get() interface which | |
3806 | * increments reference counts regardless of device state. | |
3807 | */ | |
3808 | mutex_lock(&ap->scsi_host->scan_mutex); | |
3809 | spin_lock_irqsave(ap->lock, flags); | |
3810 | ||
3811 | /* clearing dev->sdev is protected by host lock */ | |
3812 | sdev = dev->sdev; | |
3813 | dev->sdev = NULL; | |
3814 | ||
3815 | if (sdev) { | |
3816 | /* If user initiated unplug races with us, sdev can go | |
3817 | * away underneath us after the host lock and | |
3818 | * scan_mutex are released. Hold onto it. | |
3819 | */ | |
3820 | if (scsi_device_get(sdev) == 0) { | |
3821 | /* The following ensures the attached sdev is | |
3822 | * offline on return from ata_scsi_offline_dev() | |
3823 | * regardless it wins or loses the race | |
3824 | * against this function. | |
3825 | */ | |
3826 | scsi_device_set_state(sdev, SDEV_OFFLINE); | |
3827 | } else { | |
3828 | WARN_ON(1); | |
3829 | sdev = NULL; | |
3830 | } | |
3831 | } | |
3832 | ||
3833 | spin_unlock_irqrestore(ap->lock, flags); | |
3834 | mutex_unlock(&ap->scsi_host->scan_mutex); | |
3835 | ||
3836 | if (sdev) { | |
3837 | ata_dev_info(dev, "detaching (SCSI %s)\n", | |
3838 | dev_name(&sdev->sdev_gendev)); | |
3839 | ||
3840 | scsi_remove_device(sdev); | |
3841 | scsi_device_put(sdev); | |
3842 | } | |
3843 | } | |
3844 | ||
3845 | static void ata_scsi_handle_link_detach(struct ata_link *link) | |
3846 | { | |
3847 | struct ata_port *ap = link->ap; | |
3848 | struct ata_device *dev; | |
3849 | ||
3850 | ata_for_each_dev(dev, link, ALL) { | |
3851 | unsigned long flags; | |
3852 | ||
3853 | if (!(dev->flags & ATA_DFLAG_DETACHED)) | |
3854 | continue; | |
3855 | ||
3856 | spin_lock_irqsave(ap->lock, flags); | |
3857 | dev->flags &= ~ATA_DFLAG_DETACHED; | |
3858 | spin_unlock_irqrestore(ap->lock, flags); | |
3859 | ||
3860 | if (zpodd_dev_enabled(dev)) | |
3861 | zpodd_exit(dev); | |
3862 | ||
3863 | ata_scsi_remove_dev(dev); | |
3864 | } | |
3865 | } | |
3866 | ||
3867 | /** | |
3868 | * ata_scsi_media_change_notify - send media change event | |
3869 | * @dev: Pointer to the disk device with media change event | |
3870 | * | |
3871 | * Tell the block layer to send a media change notification | |
3872 | * event. | |
3873 | * | |
3874 | * LOCKING: | |
3875 | * spin_lock_irqsave(host lock) | |
3876 | */ | |
3877 | void ata_scsi_media_change_notify(struct ata_device *dev) | |
3878 | { | |
3879 | if (dev->sdev) | |
3880 | sdev_evt_send_simple(dev->sdev, SDEV_EVT_MEDIA_CHANGE, | |
3881 | GFP_ATOMIC); | |
3882 | } | |
3883 | ||
3884 | /** | |
3885 | * ata_scsi_hotplug - SCSI part of hotplug | |
3886 | * @work: Pointer to ATA port to perform SCSI hotplug on | |
3887 | * | |
3888 | * Perform SCSI part of hotplug. It's executed from a separate | |
3889 | * workqueue after EH completes. This is necessary because SCSI | |
3890 | * hot plugging requires working EH and hot unplugging is | |
3891 | * synchronized with hot plugging with a mutex. | |
3892 | * | |
3893 | * LOCKING: | |
3894 | * Kernel thread context (may sleep). | |
3895 | */ | |
3896 | void ata_scsi_hotplug(struct work_struct *work) | |
3897 | { | |
3898 | struct ata_port *ap = | |
3899 | container_of(work, struct ata_port, hotplug_task.work); | |
3900 | int i; | |
3901 | ||
3902 | if (ap->pflags & ATA_PFLAG_UNLOADING) { | |
3903 | DPRINTK("ENTER/EXIT - unloading\n"); | |
3904 | return; | |
3905 | } | |
3906 | ||
3907 | /* | |
3908 | * XXX - UGLY HACK | |
3909 | * | |
3910 | * The block layer suspend/resume path is fundamentally broken due | |
3911 | * to freezable kthreads and workqueue and may deadlock if a block | |
3912 | * device gets removed while resume is in progress. I don't know | |
3913 | * what the solution is short of removing freezable kthreads and | |
3914 | * workqueues altogether. | |
3915 | * | |
3916 | * The following is an ugly hack to avoid kicking off device | |
3917 | * removal while freezer is active. This is a joke but does avoid | |
3918 | * this particular deadlock scenario. | |
3919 | * | |
3920 | * https://bugzilla.kernel.org/show_bug.cgi?id=62801 | |
3921 | * http://marc.info/?l=linux-kernel&m=138695698516487 | |
3922 | */ | |
3923 | #ifdef CONFIG_FREEZER | |
3924 | while (pm_freezing) | |
3925 | msleep(10); | |
3926 | #endif | |
3927 | ||
3928 | DPRINTK("ENTER\n"); | |
3929 | mutex_lock(&ap->scsi_scan_mutex); | |
3930 | ||
3931 | /* Unplug detached devices. We cannot use link iterator here | |
3932 | * because PMP links have to be scanned even if PMP is | |
3933 | * currently not attached. Iterate manually. | |
3934 | */ | |
3935 | ata_scsi_handle_link_detach(&ap->link); | |
3936 | if (ap->pmp_link) | |
3937 | for (i = 0; i < SATA_PMP_MAX_PORTS; i++) | |
3938 | ata_scsi_handle_link_detach(&ap->pmp_link[i]); | |
3939 | ||
3940 | /* scan for new ones */ | |
3941 | ata_scsi_scan_host(ap, 0); | |
3942 | ||
3943 | mutex_unlock(&ap->scsi_scan_mutex); | |
3944 | DPRINTK("EXIT\n"); | |
3945 | } | |
3946 | ||
3947 | /** | |
3948 | * ata_scsi_user_scan - indication for user-initiated bus scan | |
3949 | * @shost: SCSI host to scan | |
3950 | * @channel: Channel to scan | |
3951 | * @id: ID to scan | |
3952 | * @lun: LUN to scan | |
3953 | * | |
3954 | * This function is called when user explicitly requests bus | |
3955 | * scan. Set probe pending flag and invoke EH. | |
3956 | * | |
3957 | * LOCKING: | |
3958 | * SCSI layer (we don't care) | |
3959 | * | |
3960 | * RETURNS: | |
3961 | * Zero. | |
3962 | */ | |
3963 | int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, | |
3964 | unsigned int id, u64 lun) | |
3965 | { | |
3966 | struct ata_port *ap = ata_shost_to_port(shost); | |
3967 | unsigned long flags; | |
3968 | int devno, rc = 0; | |
3969 | ||
3970 | if (!ap->ops->error_handler) | |
3971 | return -EOPNOTSUPP; | |
3972 | ||
3973 | if (lun != SCAN_WILD_CARD && lun) | |
3974 | return -EINVAL; | |
3975 | ||
3976 | if (!sata_pmp_attached(ap)) { | |
3977 | if (channel != SCAN_WILD_CARD && channel) | |
3978 | return -EINVAL; | |
3979 | devno = id; | |
3980 | } else { | |
3981 | if (id != SCAN_WILD_CARD && id) | |
3982 | return -EINVAL; | |
3983 | devno = channel; | |
3984 | } | |
3985 | ||
3986 | spin_lock_irqsave(ap->lock, flags); | |
3987 | ||
3988 | if (devno == SCAN_WILD_CARD) { | |
3989 | struct ata_link *link; | |
3990 | ||
3991 | ata_for_each_link(link, ap, EDGE) { | |
3992 | struct ata_eh_info *ehi = &link->eh_info; | |
3993 | ehi->probe_mask |= ATA_ALL_DEVICES; | |
3994 | ehi->action |= ATA_EH_RESET; | |
3995 | } | |
3996 | } else { | |
3997 | struct ata_device *dev = ata_find_dev(ap, devno); | |
3998 | ||
3999 | if (dev) { | |
4000 | struct ata_eh_info *ehi = &dev->link->eh_info; | |
4001 | ehi->probe_mask |= 1 << dev->devno; | |
4002 | ehi->action |= ATA_EH_RESET; | |
4003 | } else | |
4004 | rc = -EINVAL; | |
4005 | } | |
4006 | ||
4007 | if (rc == 0) { | |
4008 | ata_port_schedule_eh(ap); | |
4009 | spin_unlock_irqrestore(ap->lock, flags); | |
4010 | ata_port_wait_eh(ap); | |
4011 | } else | |
4012 | spin_unlock_irqrestore(ap->lock, flags); | |
4013 | ||
4014 | return rc; | |
4015 | } | |
4016 | ||
4017 | /** | |
4018 | * ata_scsi_dev_rescan - initiate scsi_rescan_device() | |
4019 | * @work: Pointer to ATA port to perform scsi_rescan_device() | |
4020 | * | |
4021 | * After ATA pass thru (SAT) commands are executed successfully, | |
4022 | * libata need to propagate the changes to SCSI layer. | |
4023 | * | |
4024 | * LOCKING: | |
4025 | * Kernel thread context (may sleep). | |
4026 | */ | |
4027 | void ata_scsi_dev_rescan(struct work_struct *work) | |
4028 | { | |
4029 | struct ata_port *ap = | |
4030 | container_of(work, struct ata_port, scsi_rescan_task); | |
4031 | struct ata_link *link; | |
4032 | struct ata_device *dev; | |
4033 | unsigned long flags; | |
4034 | ||
4035 | mutex_lock(&ap->scsi_scan_mutex); | |
4036 | spin_lock_irqsave(ap->lock, flags); | |
4037 | ||
4038 | ata_for_each_link(link, ap, EDGE) { | |
4039 | ata_for_each_dev(dev, link, ENABLED) { | |
4040 | struct scsi_device *sdev = dev->sdev; | |
4041 | ||
4042 | if (!sdev) | |
4043 | continue; | |
4044 | if (scsi_device_get(sdev)) | |
4045 | continue; | |
4046 | ||
4047 | spin_unlock_irqrestore(ap->lock, flags); | |
4048 | scsi_rescan_device(&(sdev->sdev_gendev)); | |
4049 | scsi_device_put(sdev); | |
4050 | spin_lock_irqsave(ap->lock, flags); | |
4051 | } | |
4052 | } | |
4053 | ||
4054 | spin_unlock_irqrestore(ap->lock, flags); | |
4055 | mutex_unlock(&ap->scsi_scan_mutex); | |
4056 | } | |
4057 | ||
4058 | /** | |
4059 | * ata_sas_port_alloc - Allocate port for a SAS attached SATA device | |
4060 | * @host: ATA host container for all SAS ports | |
4061 | * @port_info: Information from low-level host driver | |
4062 | * @shost: SCSI host that the scsi device is attached to | |
4063 | * | |
4064 | * LOCKING: | |
4065 | * PCI/etc. bus probe sem. | |
4066 | * | |
4067 | * RETURNS: | |
4068 | * ata_port pointer on success / NULL on failure. | |
4069 | */ | |
4070 | ||
4071 | struct ata_port *ata_sas_port_alloc(struct ata_host *host, | |
4072 | struct ata_port_info *port_info, | |
4073 | struct Scsi_Host *shost) | |
4074 | { | |
4075 | struct ata_port *ap; | |
4076 | ||
4077 | ap = ata_port_alloc(host); | |
4078 | if (!ap) | |
4079 | return NULL; | |
4080 | ||
4081 | ap->port_no = 0; | |
4082 | ap->lock = &host->lock; | |
4083 | ap->pio_mask = port_info->pio_mask; | |
4084 | ap->mwdma_mask = port_info->mwdma_mask; | |
4085 | ap->udma_mask = port_info->udma_mask; | |
4086 | ap->flags |= port_info->flags; | |
4087 | ap->ops = port_info->port_ops; | |
4088 | ap->cbl = ATA_CBL_SATA; | |
4089 | ||
4090 | return ap; | |
4091 | } | |
4092 | EXPORT_SYMBOL_GPL(ata_sas_port_alloc); | |
4093 | ||
4094 | /** | |
4095 | * ata_sas_port_start - Set port up for dma. | |
4096 | * @ap: Port to initialize | |
4097 | * | |
4098 | * Called just after data structures for each port are | |
4099 | * initialized. | |
4100 | * | |
4101 | * May be used as the port_start() entry in ata_port_operations. | |
4102 | * | |
4103 | * LOCKING: | |
4104 | * Inherited from caller. | |
4105 | */ | |
4106 | int ata_sas_port_start(struct ata_port *ap) | |
4107 | { | |
4108 | /* | |
4109 | * the port is marked as frozen at allocation time, but if we don't | |
4110 | * have new eh, we won't thaw it | |
4111 | */ | |
4112 | if (!ap->ops->error_handler) | |
4113 | ap->pflags &= ~ATA_PFLAG_FROZEN; | |
4114 | return 0; | |
4115 | } | |
4116 | EXPORT_SYMBOL_GPL(ata_sas_port_start); | |
4117 | ||
4118 | /** | |
4119 | * ata_port_stop - Undo ata_sas_port_start() | |
4120 | * @ap: Port to shut down | |
4121 | * | |
4122 | * May be used as the port_stop() entry in ata_port_operations. | |
4123 | * | |
4124 | * LOCKING: | |
4125 | * Inherited from caller. | |
4126 | */ | |
4127 | ||
4128 | void ata_sas_port_stop(struct ata_port *ap) | |
4129 | { | |
4130 | } | |
4131 | EXPORT_SYMBOL_GPL(ata_sas_port_stop); | |
4132 | ||
4133 | /** | |
4134 | * ata_sas_async_probe - simply schedule probing and return | |
4135 | * @ap: Port to probe | |
4136 | * | |
4137 | * For batch scheduling of probe for sas attached ata devices, assumes | |
4138 | * the port has already been through ata_sas_port_init() | |
4139 | */ | |
4140 | void ata_sas_async_probe(struct ata_port *ap) | |
4141 | { | |
4142 | __ata_port_probe(ap); | |
4143 | } | |
4144 | EXPORT_SYMBOL_GPL(ata_sas_async_probe); | |
4145 | ||
4146 | int ata_sas_sync_probe(struct ata_port *ap) | |
4147 | { | |
4148 | return ata_port_probe(ap); | |
4149 | } | |
4150 | EXPORT_SYMBOL_GPL(ata_sas_sync_probe); | |
4151 | ||
4152 | ||
4153 | /** | |
4154 | * ata_sas_port_init - Initialize a SATA device | |
4155 | * @ap: SATA port to initialize | |
4156 | * | |
4157 | * LOCKING: | |
4158 | * PCI/etc. bus probe sem. | |
4159 | * | |
4160 | * RETURNS: | |
4161 | * Zero on success, non-zero on error. | |
4162 | */ | |
4163 | ||
4164 | int ata_sas_port_init(struct ata_port *ap) | |
4165 | { | |
4166 | int rc = ap->ops->port_start(ap); | |
4167 | ||
4168 | if (rc) | |
4169 | return rc; | |
4170 | ap->print_id = atomic_inc_return(&ata_print_id); | |
4171 | return 0; | |
4172 | } | |
4173 | EXPORT_SYMBOL_GPL(ata_sas_port_init); | |
4174 | ||
4175 | /** | |
4176 | * ata_sas_port_destroy - Destroy a SATA port allocated by ata_sas_port_alloc | |
4177 | * @ap: SATA port to destroy | |
4178 | * | |
4179 | */ | |
4180 | ||
4181 | void ata_sas_port_destroy(struct ata_port *ap) | |
4182 | { | |
4183 | if (ap->ops->port_stop) | |
4184 | ap->ops->port_stop(ap); | |
4185 | kfree(ap); | |
4186 | } | |
4187 | EXPORT_SYMBOL_GPL(ata_sas_port_destroy); | |
4188 | ||
4189 | /** | |
4190 | * ata_sas_slave_configure - Default slave_config routine for libata devices | |
4191 | * @sdev: SCSI device to configure | |
4192 | * @ap: ATA port to which SCSI device is attached | |
4193 | * | |
4194 | * RETURNS: | |
4195 | * Zero. | |
4196 | */ | |
4197 | ||
4198 | int ata_sas_slave_configure(struct scsi_device *sdev, struct ata_port *ap) | |
4199 | { | |
4200 | ata_scsi_sdev_config(sdev); | |
4201 | ata_scsi_dev_config(sdev, ap->link.device); | |
4202 | return 0; | |
4203 | } | |
4204 | EXPORT_SYMBOL_GPL(ata_sas_slave_configure); | |
4205 | ||
4206 | /** | |
4207 | * ata_sas_queuecmd - Issue SCSI cdb to libata-managed device | |
4208 | * @cmd: SCSI command to be sent | |
4209 | * @ap: ATA port to which the command is being sent | |
4210 | * | |
4211 | * RETURNS: | |
4212 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, | |
4213 | * 0 otherwise. | |
4214 | */ | |
4215 | ||
4216 | int ata_sas_queuecmd(struct scsi_cmnd *cmd, struct ata_port *ap) | |
4217 | { | |
4218 | int rc = 0; | |
4219 | ||
4220 | ata_scsi_dump_cdb(ap, cmd); | |
4221 | ||
4222 | if (likely(ata_dev_enabled(ap->link.device))) | |
4223 | rc = __ata_scsi_queuecmd(cmd, ap->link.device); | |
4224 | else { | |
4225 | cmd->result = (DID_BAD_TARGET << 16); | |
4226 | cmd->scsi_done(cmd); | |
4227 | } | |
4228 | return rc; | |
4229 | } | |
4230 | EXPORT_SYMBOL_GPL(ata_sas_queuecmd); |