Root/
1 | /* |
2 | * libata-scsi.c - helper library for ATA |
3 | * |
4 | * Maintained by: Jeff Garzik <jgarzik@pobox.com> |
5 | * Please ALWAYS copy linux-ide@vger.kernel.org |
6 | * on emails. |
7 | * |
8 | * Copyright 2003-2004 Red Hat, Inc. All rights reserved. |
9 | * Copyright 2003-2004 Jeff Garzik |
10 | * |
11 | * |
12 | * This program is free software; you can redistribute it and/or modify |
13 | * it under the terms of the GNU General Public License as published by |
14 | * the Free Software Foundation; either version 2, or (at your option) |
15 | * any later version. |
16 | * |
17 | * This program is distributed in the hope that it will be useful, |
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
20 | * GNU General Public License for more details. |
21 | * |
22 | * You should have received a copy of the GNU General Public License |
23 | * along with this program; see the file COPYING. If not, write to |
24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. |
25 | * |
26 | * |
27 | * libata documentation is available via 'make {ps|pdf}docs', |
28 | * as Documentation/DocBook/libata.* |
29 | * |
30 | * Hardware documentation available from |
31 | * - http://www.t10.org/ |
32 | * - http://www.t13.org/ |
33 | * |
34 | */ |
35 | |
36 | #include <linux/slab.h> |
37 | #include <linux/kernel.h> |
38 | #include <linux/blkdev.h> |
39 | #include <linux/spinlock.h> |
40 | #include <linux/export.h> |
41 | #include <scsi/scsi.h> |
42 | #include <scsi/scsi_host.h> |
43 | #include <scsi/scsi_cmnd.h> |
44 | #include <scsi/scsi_eh.h> |
45 | #include <scsi/scsi_device.h> |
46 | #include <scsi/scsi_tcq.h> |
47 | #include <scsi/scsi_transport.h> |
48 | #include <linux/libata.h> |
49 | #include <linux/hdreg.h> |
50 | #include <linux/uaccess.h> |
51 | #include <linux/suspend.h> |
52 | #include <asm/unaligned.h> |
53 | |
54 | #include "libata.h" |
55 | #include "libata-transport.h" |
56 | |
57 | #define ATA_SCSI_RBUF_SIZE 4096 |
58 | |
59 | static DEFINE_SPINLOCK(ata_scsi_rbuf_lock); |
60 | static u8 ata_scsi_rbuf[ATA_SCSI_RBUF_SIZE]; |
61 | |
62 | typedef unsigned int (*ata_xlat_func_t)(struct ata_queued_cmd *qc); |
63 | |
64 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, |
65 | const struct scsi_device *scsidev); |
66 | static struct ata_device *ata_scsi_find_dev(struct ata_port *ap, |
67 | const struct scsi_device *scsidev); |
68 | |
69 | #define RW_RECOVERY_MPAGE 0x1 |
70 | #define RW_RECOVERY_MPAGE_LEN 12 |
71 | #define CACHE_MPAGE 0x8 |
72 | #define CACHE_MPAGE_LEN 20 |
73 | #define CONTROL_MPAGE 0xa |
74 | #define CONTROL_MPAGE_LEN 12 |
75 | #define ALL_MPAGES 0x3f |
76 | #define ALL_SUB_MPAGES 0xff |
77 | |
78 | |
79 | static const u8 def_rw_recovery_mpage[RW_RECOVERY_MPAGE_LEN] = { |
80 | RW_RECOVERY_MPAGE, |
81 | RW_RECOVERY_MPAGE_LEN - 2, |
82 | (1 << 7), /* AWRE */ |
83 | 0, /* read retry count */ |
84 | 0, 0, 0, 0, |
85 | 0, /* write retry count */ |
86 | 0, 0, 0 |
87 | }; |
88 | |
89 | static const u8 def_cache_mpage[CACHE_MPAGE_LEN] = { |
90 | CACHE_MPAGE, |
91 | CACHE_MPAGE_LEN - 2, |
92 | 0, /* contains WCE, needs to be 0 for logic */ |
93 | 0, 0, 0, 0, 0, 0, 0, 0, 0, |
94 | 0, /* contains DRA, needs to be 0 for logic */ |
95 | 0, 0, 0, 0, 0, 0, 0 |
96 | }; |
97 | |
98 | static const u8 def_control_mpage[CONTROL_MPAGE_LEN] = { |
99 | CONTROL_MPAGE, |
100 | CONTROL_MPAGE_LEN - 2, |
101 | 2, /* DSENSE=0, GLTSD=1 */ |
102 | 0, /* [QAM+QERR may be 1, see 05-359r1] */ |
103 | 0, 0, 0, 0, 0xff, 0xff, |
104 | 0, 30 /* extended self test time, see 05-359r1 */ |
105 | }; |
106 | |
107 | static const char *ata_lpm_policy_names[] = { |
108 | [ATA_LPM_UNKNOWN] = "max_performance", |
109 | [ATA_LPM_MAX_POWER] = "max_performance", |
110 | [ATA_LPM_MED_POWER] = "medium_power", |
111 | [ATA_LPM_MIN_POWER] = "min_power", |
112 | }; |
113 | |
114 | static ssize_t ata_scsi_lpm_store(struct device *dev, |
115 | struct device_attribute *attr, |
116 | const char *buf, size_t count) |
117 | { |
118 | struct Scsi_Host *shost = class_to_shost(dev); |
119 | struct ata_port *ap = ata_shost_to_port(shost); |
120 | enum ata_lpm_policy policy; |
121 | unsigned long flags; |
122 | |
123 | /* UNKNOWN is internal state, iterate from MAX_POWER */ |
124 | for (policy = ATA_LPM_MAX_POWER; |
125 | policy < ARRAY_SIZE(ata_lpm_policy_names); policy++) { |
126 | const char *name = ata_lpm_policy_names[policy]; |
127 | |
128 | if (strncmp(name, buf, strlen(name)) == 0) |
129 | break; |
130 | } |
131 | if (policy == ARRAY_SIZE(ata_lpm_policy_names)) |
132 | return -EINVAL; |
133 | |
134 | spin_lock_irqsave(ap->lock, flags); |
135 | ap->target_lpm_policy = policy; |
136 | ata_port_schedule_eh(ap); |
137 | spin_unlock_irqrestore(ap->lock, flags); |
138 | |
139 | return count; |
140 | } |
141 | |
142 | static ssize_t ata_scsi_lpm_show(struct device *dev, |
143 | struct device_attribute *attr, char *buf) |
144 | { |
145 | struct Scsi_Host *shost = class_to_shost(dev); |
146 | struct ata_port *ap = ata_shost_to_port(shost); |
147 | |
148 | if (ap->target_lpm_policy >= ARRAY_SIZE(ata_lpm_policy_names)) |
149 | return -EINVAL; |
150 | |
151 | return snprintf(buf, PAGE_SIZE, "%s\n", |
152 | ata_lpm_policy_names[ap->target_lpm_policy]); |
153 | } |
154 | DEVICE_ATTR(link_power_management_policy, S_IRUGO | S_IWUSR, |
155 | ata_scsi_lpm_show, ata_scsi_lpm_store); |
156 | EXPORT_SYMBOL_GPL(dev_attr_link_power_management_policy); |
157 | |
158 | static ssize_t ata_scsi_park_show(struct device *device, |
159 | struct device_attribute *attr, char *buf) |
160 | { |
161 | struct scsi_device *sdev = to_scsi_device(device); |
162 | struct ata_port *ap; |
163 | struct ata_link *link; |
164 | struct ata_device *dev; |
165 | unsigned long flags, now; |
166 | unsigned int uninitialized_var(msecs); |
167 | int rc = 0; |
168 | |
169 | ap = ata_shost_to_port(sdev->host); |
170 | |
171 | spin_lock_irqsave(ap->lock, flags); |
172 | dev = ata_scsi_find_dev(ap, sdev); |
173 | if (!dev) { |
174 | rc = -ENODEV; |
175 | goto unlock; |
176 | } |
177 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { |
178 | rc = -EOPNOTSUPP; |
179 | goto unlock; |
180 | } |
181 | |
182 | link = dev->link; |
183 | now = jiffies; |
184 | if (ap->pflags & ATA_PFLAG_EH_IN_PROGRESS && |
185 | link->eh_context.unloaded_mask & (1 << dev->devno) && |
186 | time_after(dev->unpark_deadline, now)) |
187 | msecs = jiffies_to_msecs(dev->unpark_deadline - now); |
188 | else |
189 | msecs = 0; |
190 | |
191 | unlock: |
192 | spin_unlock_irq(ap->lock); |
193 | |
194 | return rc ? rc : snprintf(buf, 20, "%u\n", msecs); |
195 | } |
196 | |
197 | static ssize_t ata_scsi_park_store(struct device *device, |
198 | struct device_attribute *attr, |
199 | const char *buf, size_t len) |
200 | { |
201 | struct scsi_device *sdev = to_scsi_device(device); |
202 | struct ata_port *ap; |
203 | struct ata_device *dev; |
204 | long int input; |
205 | unsigned long flags; |
206 | int rc; |
207 | |
208 | rc = strict_strtol(buf, 10, &input); |
209 | if (rc || input < -2) |
210 | return -EINVAL; |
211 | if (input > ATA_TMOUT_MAX_PARK) { |
212 | rc = -EOVERFLOW; |
213 | input = ATA_TMOUT_MAX_PARK; |
214 | } |
215 | |
216 | ap = ata_shost_to_port(sdev->host); |
217 | |
218 | spin_lock_irqsave(ap->lock, flags); |
219 | dev = ata_scsi_find_dev(ap, sdev); |
220 | if (unlikely(!dev)) { |
221 | rc = -ENODEV; |
222 | goto unlock; |
223 | } |
224 | if (dev->class != ATA_DEV_ATA) { |
225 | rc = -EOPNOTSUPP; |
226 | goto unlock; |
227 | } |
228 | |
229 | if (input >= 0) { |
230 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { |
231 | rc = -EOPNOTSUPP; |
232 | goto unlock; |
233 | } |
234 | |
235 | dev->unpark_deadline = ata_deadline(jiffies, input); |
236 | dev->link->eh_info.dev_action[dev->devno] |= ATA_EH_PARK; |
237 | ata_port_schedule_eh(ap); |
238 | complete(&ap->park_req_pending); |
239 | } else { |
240 | switch (input) { |
241 | case -1: |
242 | dev->flags &= ~ATA_DFLAG_NO_UNLOAD; |
243 | break; |
244 | case -2: |
245 | dev->flags |= ATA_DFLAG_NO_UNLOAD; |
246 | break; |
247 | } |
248 | } |
249 | unlock: |
250 | spin_unlock_irqrestore(ap->lock, flags); |
251 | |
252 | return rc ? rc : len; |
253 | } |
254 | DEVICE_ATTR(unload_heads, S_IRUGO | S_IWUSR, |
255 | ata_scsi_park_show, ata_scsi_park_store); |
256 | EXPORT_SYMBOL_GPL(dev_attr_unload_heads); |
257 | |
258 | static void ata_scsi_set_sense(struct scsi_cmnd *cmd, u8 sk, u8 asc, u8 ascq) |
259 | { |
260 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
261 | |
262 | scsi_build_sense_buffer(0, cmd->sense_buffer, sk, asc, ascq); |
263 | } |
264 | |
265 | static ssize_t |
266 | ata_scsi_em_message_store(struct device *dev, struct device_attribute *attr, |
267 | const char *buf, size_t count) |
268 | { |
269 | struct Scsi_Host *shost = class_to_shost(dev); |
270 | struct ata_port *ap = ata_shost_to_port(shost); |
271 | if (ap->ops->em_store && (ap->flags & ATA_FLAG_EM)) |
272 | return ap->ops->em_store(ap, buf, count); |
273 | return -EINVAL; |
274 | } |
275 | |
276 | static ssize_t |
277 | ata_scsi_em_message_show(struct device *dev, struct device_attribute *attr, |
278 | char *buf) |
279 | { |
280 | struct Scsi_Host *shost = class_to_shost(dev); |
281 | struct ata_port *ap = ata_shost_to_port(shost); |
282 | |
283 | if (ap->ops->em_show && (ap->flags & ATA_FLAG_EM)) |
284 | return ap->ops->em_show(ap, buf); |
285 | return -EINVAL; |
286 | } |
287 | DEVICE_ATTR(em_message, S_IRUGO | S_IWUSR, |
288 | ata_scsi_em_message_show, ata_scsi_em_message_store); |
289 | EXPORT_SYMBOL_GPL(dev_attr_em_message); |
290 | |
291 | static ssize_t |
292 | ata_scsi_em_message_type_show(struct device *dev, struct device_attribute *attr, |
293 | char *buf) |
294 | { |
295 | struct Scsi_Host *shost = class_to_shost(dev); |
296 | struct ata_port *ap = ata_shost_to_port(shost); |
297 | |
298 | return snprintf(buf, 23, "%d\n", ap->em_message_type); |
299 | } |
300 | DEVICE_ATTR(em_message_type, S_IRUGO, |
301 | ata_scsi_em_message_type_show, NULL); |
302 | EXPORT_SYMBOL_GPL(dev_attr_em_message_type); |
303 | |
304 | static ssize_t |
305 | ata_scsi_activity_show(struct device *dev, struct device_attribute *attr, |
306 | char *buf) |
307 | { |
308 | struct scsi_device *sdev = to_scsi_device(dev); |
309 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
310 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); |
311 | |
312 | if (atadev && ap->ops->sw_activity_show && |
313 | (ap->flags & ATA_FLAG_SW_ACTIVITY)) |
314 | return ap->ops->sw_activity_show(atadev, buf); |
315 | return -EINVAL; |
316 | } |
317 | |
318 | static ssize_t |
319 | ata_scsi_activity_store(struct device *dev, struct device_attribute *attr, |
320 | const char *buf, size_t count) |
321 | { |
322 | struct scsi_device *sdev = to_scsi_device(dev); |
323 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
324 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); |
325 | enum sw_activity val; |
326 | int rc; |
327 | |
328 | if (atadev && ap->ops->sw_activity_store && |
329 | (ap->flags & ATA_FLAG_SW_ACTIVITY)) { |
330 | val = simple_strtoul(buf, NULL, 0); |
331 | switch (val) { |
332 | case OFF: case BLINK_ON: case BLINK_OFF: |
333 | rc = ap->ops->sw_activity_store(atadev, val); |
334 | if (!rc) |
335 | return count; |
336 | else |
337 | return rc; |
338 | } |
339 | } |
340 | return -EINVAL; |
341 | } |
342 | DEVICE_ATTR(sw_activity, S_IWUSR | S_IRUGO, ata_scsi_activity_show, |
343 | ata_scsi_activity_store); |
344 | EXPORT_SYMBOL_GPL(dev_attr_sw_activity); |
345 | |
346 | struct device_attribute *ata_common_sdev_attrs[] = { |
347 | &dev_attr_unload_heads, |
348 | NULL |
349 | }; |
350 | EXPORT_SYMBOL_GPL(ata_common_sdev_attrs); |
351 | |
352 | static void ata_scsi_invalid_field(struct scsi_cmnd *cmd) |
353 | { |
354 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x24, 0x0); |
355 | /* "Invalid field in cbd" */ |
356 | cmd->scsi_done(cmd); |
357 | } |
358 | |
359 | /** |
360 | * ata_std_bios_param - generic bios head/sector/cylinder calculator used by sd. |
361 | * @sdev: SCSI device for which BIOS geometry is to be determined |
362 | * @bdev: block device associated with @sdev |
363 | * @capacity: capacity of SCSI device |
364 | * @geom: location to which geometry will be output |
365 | * |
366 | * Generic bios head/sector/cylinder calculator |
367 | * used by sd. Most BIOSes nowadays expect a XXX/255/16 (CHS) |
368 | * mapping. Some situations may arise where the disk is not |
369 | * bootable if this is not used. |
370 | * |
371 | * LOCKING: |
372 | * Defined by the SCSI layer. We don't really care. |
373 | * |
374 | * RETURNS: |
375 | * Zero. |
376 | */ |
377 | int ata_std_bios_param(struct scsi_device *sdev, struct block_device *bdev, |
378 | sector_t capacity, int geom[]) |
379 | { |
380 | geom[0] = 255; |
381 | geom[1] = 63; |
382 | sector_div(capacity, 255*63); |
383 | geom[2] = capacity; |
384 | |
385 | return 0; |
386 | } |
387 | |
388 | /** |
389 | * ata_scsi_unlock_native_capacity - unlock native capacity |
390 | * @sdev: SCSI device to adjust device capacity for |
391 | * |
392 | * This function is called if a partition on @sdev extends beyond |
393 | * the end of the device. It requests EH to unlock HPA. |
394 | * |
395 | * LOCKING: |
396 | * Defined by the SCSI layer. Might sleep. |
397 | */ |
398 | void ata_scsi_unlock_native_capacity(struct scsi_device *sdev) |
399 | { |
400 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
401 | struct ata_device *dev; |
402 | unsigned long flags; |
403 | |
404 | spin_lock_irqsave(ap->lock, flags); |
405 | |
406 | dev = ata_scsi_find_dev(ap, sdev); |
407 | if (dev && dev->n_sectors < dev->n_native_sectors) { |
408 | dev->flags |= ATA_DFLAG_UNLOCK_HPA; |
409 | dev->link->eh_info.action |= ATA_EH_RESET; |
410 | ata_port_schedule_eh(ap); |
411 | } |
412 | |
413 | spin_unlock_irqrestore(ap->lock, flags); |
414 | ata_port_wait_eh(ap); |
415 | } |
416 | |
417 | /** |
418 | * ata_get_identity - Handler for HDIO_GET_IDENTITY ioctl |
419 | * @ap: target port |
420 | * @sdev: SCSI device to get identify data for |
421 | * @arg: User buffer area for identify data |
422 | * |
423 | * LOCKING: |
424 | * Defined by the SCSI layer. We don't really care. |
425 | * |
426 | * RETURNS: |
427 | * Zero on success, negative errno on error. |
428 | */ |
429 | static int ata_get_identity(struct ata_port *ap, struct scsi_device *sdev, |
430 | void __user *arg) |
431 | { |
432 | struct ata_device *dev = ata_scsi_find_dev(ap, sdev); |
433 | u16 __user *dst = arg; |
434 | char buf[40]; |
435 | |
436 | if (!dev) |
437 | return -ENOMSG; |
438 | |
439 | if (copy_to_user(dst, dev->id, ATA_ID_WORDS * sizeof(u16))) |
440 | return -EFAULT; |
441 | |
442 | ata_id_string(dev->id, buf, ATA_ID_PROD, ATA_ID_PROD_LEN); |
443 | if (copy_to_user(dst + ATA_ID_PROD, buf, ATA_ID_PROD_LEN)) |
444 | return -EFAULT; |
445 | |
446 | ata_id_string(dev->id, buf, ATA_ID_FW_REV, ATA_ID_FW_REV_LEN); |
447 | if (copy_to_user(dst + ATA_ID_FW_REV, buf, ATA_ID_FW_REV_LEN)) |
448 | return -EFAULT; |
449 | |
450 | ata_id_string(dev->id, buf, ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
451 | if (copy_to_user(dst + ATA_ID_SERNO, buf, ATA_ID_SERNO_LEN)) |
452 | return -EFAULT; |
453 | |
454 | return 0; |
455 | } |
456 | |
457 | /** |
458 | * ata_cmd_ioctl - Handler for HDIO_DRIVE_CMD ioctl |
459 | * @scsidev: Device to which we are issuing command |
460 | * @arg: User provided data for issuing command |
461 | * |
462 | * LOCKING: |
463 | * Defined by the SCSI layer. We don't really care. |
464 | * |
465 | * RETURNS: |
466 | * Zero on success, negative errno on error. |
467 | */ |
468 | int ata_cmd_ioctl(struct scsi_device *scsidev, void __user *arg) |
469 | { |
470 | int rc = 0; |
471 | u8 scsi_cmd[MAX_COMMAND_SIZE]; |
472 | u8 args[4], *argbuf = NULL, *sensebuf = NULL; |
473 | int argsize = 0; |
474 | enum dma_data_direction data_dir; |
475 | int cmd_result; |
476 | |
477 | if (arg == NULL) |
478 | return -EINVAL; |
479 | |
480 | if (copy_from_user(args, arg, sizeof(args))) |
481 | return -EFAULT; |
482 | |
483 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); |
484 | if (!sensebuf) |
485 | return -ENOMEM; |
486 | |
487 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); |
488 | |
489 | if (args[3]) { |
490 | argsize = ATA_SECT_SIZE * args[3]; |
491 | argbuf = kmalloc(argsize, GFP_KERNEL); |
492 | if (argbuf == NULL) { |
493 | rc = -ENOMEM; |
494 | goto error; |
495 | } |
496 | |
497 | scsi_cmd[1] = (4 << 1); /* PIO Data-in */ |
498 | scsi_cmd[2] = 0x0e; /* no off.line or cc, read from dev, |
499 | block count in sector count field */ |
500 | data_dir = DMA_FROM_DEVICE; |
501 | } else { |
502 | scsi_cmd[1] = (3 << 1); /* Non-data */ |
503 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ |
504 | data_dir = DMA_NONE; |
505 | } |
506 | |
507 | scsi_cmd[0] = ATA_16; |
508 | |
509 | scsi_cmd[4] = args[2]; |
510 | if (args[0] == ATA_CMD_SMART) { /* hack -- ide driver does this too */ |
511 | scsi_cmd[6] = args[3]; |
512 | scsi_cmd[8] = args[1]; |
513 | scsi_cmd[10] = 0x4f; |
514 | scsi_cmd[12] = 0xc2; |
515 | } else { |
516 | scsi_cmd[6] = args[1]; |
517 | } |
518 | scsi_cmd[14] = args[0]; |
519 | |
520 | /* Good values for timeout and retries? Values below |
521 | from scsi_ioctl_send_command() for default case... */ |
522 | cmd_result = scsi_execute(scsidev, scsi_cmd, data_dir, argbuf, argsize, |
523 | sensebuf, (10*HZ), 5, 0, NULL); |
524 | |
525 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ |
526 | u8 *desc = sensebuf + 8; |
527 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ |
528 | |
529 | /* If we set cc then ATA pass-through will cause a |
530 | * check condition even if no error. Filter that. */ |
531 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { |
532 | struct scsi_sense_hdr sshdr; |
533 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, |
534 | &sshdr); |
535 | if (sshdr.sense_key == RECOVERED_ERROR && |
536 | sshdr.asc == 0 && sshdr.ascq == 0x1d) |
537 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; |
538 | } |
539 | |
540 | /* Send userspace a few ATA registers (same as drivers/ide) */ |
541 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ |
542 | desc[0] == 0x09) { /* code is "ATA Descriptor" */ |
543 | args[0] = desc[13]; /* status */ |
544 | args[1] = desc[3]; /* error */ |
545 | args[2] = desc[5]; /* sector count (0:7) */ |
546 | if (copy_to_user(arg, args, sizeof(args))) |
547 | rc = -EFAULT; |
548 | } |
549 | } |
550 | |
551 | |
552 | if (cmd_result) { |
553 | rc = -EIO; |
554 | goto error; |
555 | } |
556 | |
557 | if ((argbuf) |
558 | && copy_to_user(arg + sizeof(args), argbuf, argsize)) |
559 | rc = -EFAULT; |
560 | error: |
561 | kfree(sensebuf); |
562 | kfree(argbuf); |
563 | return rc; |
564 | } |
565 | |
566 | /** |
567 | * ata_task_ioctl - Handler for HDIO_DRIVE_TASK ioctl |
568 | * @scsidev: Device to which we are issuing command |
569 | * @arg: User provided data for issuing command |
570 | * |
571 | * LOCKING: |
572 | * Defined by the SCSI layer. We don't really care. |
573 | * |
574 | * RETURNS: |
575 | * Zero on success, negative errno on error. |
576 | */ |
577 | int ata_task_ioctl(struct scsi_device *scsidev, void __user *arg) |
578 | { |
579 | int rc = 0; |
580 | u8 scsi_cmd[MAX_COMMAND_SIZE]; |
581 | u8 args[7], *sensebuf = NULL; |
582 | int cmd_result; |
583 | |
584 | if (arg == NULL) |
585 | return -EINVAL; |
586 | |
587 | if (copy_from_user(args, arg, sizeof(args))) |
588 | return -EFAULT; |
589 | |
590 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); |
591 | if (!sensebuf) |
592 | return -ENOMEM; |
593 | |
594 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); |
595 | scsi_cmd[0] = ATA_16; |
596 | scsi_cmd[1] = (3 << 1); /* Non-data */ |
597 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ |
598 | scsi_cmd[4] = args[1]; |
599 | scsi_cmd[6] = args[2]; |
600 | scsi_cmd[8] = args[3]; |
601 | scsi_cmd[10] = args[4]; |
602 | scsi_cmd[12] = args[5]; |
603 | scsi_cmd[13] = args[6] & 0x4f; |
604 | scsi_cmd[14] = args[0]; |
605 | |
606 | /* Good values for timeout and retries? Values below |
607 | from scsi_ioctl_send_command() for default case... */ |
608 | cmd_result = scsi_execute(scsidev, scsi_cmd, DMA_NONE, NULL, 0, |
609 | sensebuf, (10*HZ), 5, 0, NULL); |
610 | |
611 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ |
612 | u8 *desc = sensebuf + 8; |
613 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ |
614 | |
615 | /* If we set cc then ATA pass-through will cause a |
616 | * check condition even if no error. Filter that. */ |
617 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { |
618 | struct scsi_sense_hdr sshdr; |
619 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, |
620 | &sshdr); |
621 | if (sshdr.sense_key == RECOVERED_ERROR && |
622 | sshdr.asc == 0 && sshdr.ascq == 0x1d) |
623 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; |
624 | } |
625 | |
626 | /* Send userspace ATA registers */ |
627 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ |
628 | desc[0] == 0x09) {/* code is "ATA Descriptor" */ |
629 | args[0] = desc[13]; /* status */ |
630 | args[1] = desc[3]; /* error */ |
631 | args[2] = desc[5]; /* sector count (0:7) */ |
632 | args[3] = desc[7]; /* lbal */ |
633 | args[4] = desc[9]; /* lbam */ |
634 | args[5] = desc[11]; /* lbah */ |
635 | args[6] = desc[12]; /* select */ |
636 | if (copy_to_user(arg, args, sizeof(args))) |
637 | rc = -EFAULT; |
638 | } |
639 | } |
640 | |
641 | if (cmd_result) { |
642 | rc = -EIO; |
643 | goto error; |
644 | } |
645 | |
646 | error: |
647 | kfree(sensebuf); |
648 | return rc; |
649 | } |
650 | |
651 | static int ata_ioc32(struct ata_port *ap) |
652 | { |
653 | if (ap->flags & ATA_FLAG_PIO_DMA) |
654 | return 1; |
655 | if (ap->pflags & ATA_PFLAG_PIO32) |
656 | return 1; |
657 | return 0; |
658 | } |
659 | |
660 | int ata_sas_scsi_ioctl(struct ata_port *ap, struct scsi_device *scsidev, |
661 | int cmd, void __user *arg) |
662 | { |
663 | int val = -EINVAL, rc = -EINVAL; |
664 | unsigned long flags; |
665 | |
666 | switch (cmd) { |
667 | case ATA_IOC_GET_IO32: |
668 | spin_lock_irqsave(ap->lock, flags); |
669 | val = ata_ioc32(ap); |
670 | spin_unlock_irqrestore(ap->lock, flags); |
671 | if (copy_to_user(arg, &val, 1)) |
672 | return -EFAULT; |
673 | return 0; |
674 | |
675 | case ATA_IOC_SET_IO32: |
676 | val = (unsigned long) arg; |
677 | rc = 0; |
678 | spin_lock_irqsave(ap->lock, flags); |
679 | if (ap->pflags & ATA_PFLAG_PIO32CHANGE) { |
680 | if (val) |
681 | ap->pflags |= ATA_PFLAG_PIO32; |
682 | else |
683 | ap->pflags &= ~ATA_PFLAG_PIO32; |
684 | } else { |
685 | if (val != ata_ioc32(ap)) |
686 | rc = -EINVAL; |
687 | } |
688 | spin_unlock_irqrestore(ap->lock, flags); |
689 | return rc; |
690 | |
691 | case HDIO_GET_IDENTITY: |
692 | return ata_get_identity(ap, scsidev, arg); |
693 | |
694 | case HDIO_DRIVE_CMD: |
695 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) |
696 | return -EACCES; |
697 | return ata_cmd_ioctl(scsidev, arg); |
698 | |
699 | case HDIO_DRIVE_TASK: |
700 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) |
701 | return -EACCES; |
702 | return ata_task_ioctl(scsidev, arg); |
703 | |
704 | default: |
705 | rc = -ENOTTY; |
706 | break; |
707 | } |
708 | |
709 | return rc; |
710 | } |
711 | EXPORT_SYMBOL_GPL(ata_sas_scsi_ioctl); |
712 | |
713 | int ata_scsi_ioctl(struct scsi_device *scsidev, int cmd, void __user *arg) |
714 | { |
715 | return ata_sas_scsi_ioctl(ata_shost_to_port(scsidev->host), |
716 | scsidev, cmd, arg); |
717 | } |
718 | EXPORT_SYMBOL_GPL(ata_scsi_ioctl); |
719 | |
720 | /** |
721 | * ata_scsi_qc_new - acquire new ata_queued_cmd reference |
722 | * @dev: ATA device to which the new command is attached |
723 | * @cmd: SCSI command that originated this ATA command |
724 | * |
725 | * Obtain a reference to an unused ata_queued_cmd structure, |
726 | * which is the basic libata structure representing a single |
727 | * ATA command sent to the hardware. |
728 | * |
729 | * If a command was available, fill in the SCSI-specific |
730 | * portions of the structure with information on the |
731 | * current command. |
732 | * |
733 | * LOCKING: |
734 | * spin_lock_irqsave(host lock) |
735 | * |
736 | * RETURNS: |
737 | * Command allocated, or %NULL if none available. |
738 | */ |
739 | static struct ata_queued_cmd *ata_scsi_qc_new(struct ata_device *dev, |
740 | struct scsi_cmnd *cmd) |
741 | { |
742 | struct ata_queued_cmd *qc; |
743 | |
744 | qc = ata_qc_new_init(dev); |
745 | if (qc) { |
746 | qc->scsicmd = cmd; |
747 | qc->scsidone = cmd->scsi_done; |
748 | |
749 | qc->sg = scsi_sglist(cmd); |
750 | qc->n_elem = scsi_sg_count(cmd); |
751 | } else { |
752 | cmd->result = (DID_OK << 16) | (QUEUE_FULL << 1); |
753 | cmd->scsi_done(cmd); |
754 | } |
755 | |
756 | return qc; |
757 | } |
758 | |
759 | static void ata_qc_set_pc_nbytes(struct ata_queued_cmd *qc) |
760 | { |
761 | struct scsi_cmnd *scmd = qc->scsicmd; |
762 | |
763 | qc->extrabytes = scmd->request->extra_len; |
764 | qc->nbytes = scsi_bufflen(scmd) + qc->extrabytes; |
765 | } |
766 | |
767 | /** |
768 | * ata_dump_status - user friendly display of error info |
769 | * @id: id of the port in question |
770 | * @tf: ptr to filled out taskfile |
771 | * |
772 | * Decode and dump the ATA error/status registers for the user so |
773 | * that they have some idea what really happened at the non |
774 | * make-believe layer. |
775 | * |
776 | * LOCKING: |
777 | * inherited from caller |
778 | */ |
779 | static void ata_dump_status(unsigned id, struct ata_taskfile *tf) |
780 | { |
781 | u8 stat = tf->command, err = tf->feature; |
782 | |
783 | printk(KERN_WARNING "ata%u: status=0x%02x { ", id, stat); |
784 | if (stat & ATA_BUSY) { |
785 | printk("Busy }\n"); /* Data is not valid in this case */ |
786 | } else { |
787 | if (stat & 0x40) printk("DriveReady "); |
788 | if (stat & 0x20) printk("DeviceFault "); |
789 | if (stat & 0x10) printk("SeekComplete "); |
790 | if (stat & 0x08) printk("DataRequest "); |
791 | if (stat & 0x04) printk("CorrectedError "); |
792 | if (stat & 0x02) printk("Index "); |
793 | if (stat & 0x01) printk("Error "); |
794 | printk("}\n"); |
795 | |
796 | if (err) { |
797 | printk(KERN_WARNING "ata%u: error=0x%02x { ", id, err); |
798 | if (err & 0x04) printk("DriveStatusError "); |
799 | if (err & 0x80) { |
800 | if (err & 0x04) printk("BadCRC "); |
801 | else printk("Sector "); |
802 | } |
803 | if (err & 0x40) printk("UncorrectableError "); |
804 | if (err & 0x10) printk("SectorIdNotFound "); |
805 | if (err & 0x02) printk("TrackZeroNotFound "); |
806 | if (err & 0x01) printk("AddrMarkNotFound "); |
807 | printk("}\n"); |
808 | } |
809 | } |
810 | } |
811 | |
812 | /** |
813 | * ata_to_sense_error - convert ATA error to SCSI error |
814 | * @id: ATA device number |
815 | * @drv_stat: value contained in ATA status register |
816 | * @drv_err: value contained in ATA error register |
817 | * @sk: the sense key we'll fill out |
818 | * @asc: the additional sense code we'll fill out |
819 | * @ascq: the additional sense code qualifier we'll fill out |
820 | * @verbose: be verbose |
821 | * |
822 | * Converts an ATA error into a SCSI error. Fill out pointers to |
823 | * SK, ASC, and ASCQ bytes for later use in fixed or descriptor |
824 | * format sense blocks. |
825 | * |
826 | * LOCKING: |
827 | * spin_lock_irqsave(host lock) |
828 | */ |
829 | static void ata_to_sense_error(unsigned id, u8 drv_stat, u8 drv_err, u8 *sk, |
830 | u8 *asc, u8 *ascq, int verbose) |
831 | { |
832 | int i; |
833 | |
834 | /* Based on the 3ware driver translation table */ |
835 | static const unsigned char sense_table[][4] = { |
836 | /* BBD|ECC|ID|MAR */ |
837 | {0xd1, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command |
838 | /* BBD|ECC|ID */ |
839 | {0xd0, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command |
840 | /* ECC|MC|MARK */ |
841 | {0x61, HARDWARE_ERROR, 0x00, 0x00}, // Device fault Hardware error |
842 | /* ICRC|ABRT */ /* NB: ICRC & !ABRT is BBD */ |
843 | {0x84, ABORTED_COMMAND, 0x47, 0x00}, // Data CRC error SCSI parity error |
844 | /* MC|ID|ABRT|TRK0|MARK */ |
845 | {0x37, NOT_READY, 0x04, 0x00}, // Unit offline Not ready |
846 | /* MCR|MARK */ |
847 | {0x09, NOT_READY, 0x04, 0x00}, // Unrecovered disk error Not ready |
848 | /* Bad address mark */ |
849 | {0x01, MEDIUM_ERROR, 0x13, 0x00}, // Address mark not found Address mark not found for data field |
850 | /* TRK0 */ |
851 | {0x02, HARDWARE_ERROR, 0x00, 0x00}, // Track 0 not found Hardware error |
852 | /* Abort & !ICRC */ |
853 | {0x04, ABORTED_COMMAND, 0x00, 0x00}, // Aborted command Aborted command |
854 | /* Media change request */ |
855 | {0x08, NOT_READY, 0x04, 0x00}, // Media change request FIXME: faking offline |
856 | /* SRV */ |
857 | {0x10, ABORTED_COMMAND, 0x14, 0x00}, // ID not found Recorded entity not found |
858 | /* Media change */ |
859 | {0x08, NOT_READY, 0x04, 0x00}, // Media change FIXME: faking offline |
860 | /* ECC */ |
861 | {0x40, MEDIUM_ERROR, 0x11, 0x04}, // Uncorrectable ECC error Unrecovered read error |
862 | /* BBD - block marked bad */ |
863 | {0x80, MEDIUM_ERROR, 0x11, 0x04}, // Block marked bad Medium error, unrecovered read error |
864 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark |
865 | }; |
866 | static const unsigned char stat_table[][4] = { |
867 | /* Must be first because BUSY means no other bits valid */ |
868 | {0x80, ABORTED_COMMAND, 0x47, 0x00}, // Busy, fake parity for now |
869 | {0x20, HARDWARE_ERROR, 0x00, 0x00}, // Device fault |
870 | {0x08, ABORTED_COMMAND, 0x47, 0x00}, // Timed out in xfer, fake parity for now |
871 | {0x04, RECOVERED_ERROR, 0x11, 0x00}, // Recovered ECC error Medium error, recovered |
872 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark |
873 | }; |
874 | |
875 | /* |
876 | * Is this an error we can process/parse |
877 | */ |
878 | if (drv_stat & ATA_BUSY) { |
879 | drv_err = 0; /* Ignore the err bits, they're invalid */ |
880 | } |
881 | |
882 | if (drv_err) { |
883 | /* Look for drv_err */ |
884 | for (i = 0; sense_table[i][0] != 0xFF; i++) { |
885 | /* Look for best matches first */ |
886 | if ((sense_table[i][0] & drv_err) == |
887 | sense_table[i][0]) { |
888 | *sk = sense_table[i][1]; |
889 | *asc = sense_table[i][2]; |
890 | *ascq = sense_table[i][3]; |
891 | goto translate_done; |
892 | } |
893 | } |
894 | /* No immediate match */ |
895 | if (verbose) |
896 | printk(KERN_WARNING "ata%u: no sense translation for " |
897 | "error 0x%02x\n", id, drv_err); |
898 | } |
899 | |
900 | /* Fall back to interpreting status bits */ |
901 | for (i = 0; stat_table[i][0] != 0xFF; i++) { |
902 | if (stat_table[i][0] & drv_stat) { |
903 | *sk = stat_table[i][1]; |
904 | *asc = stat_table[i][2]; |
905 | *ascq = stat_table[i][3]; |
906 | goto translate_done; |
907 | } |
908 | } |
909 | /* No error? Undecoded? */ |
910 | if (verbose) |
911 | printk(KERN_WARNING "ata%u: no sense translation for " |
912 | "status: 0x%02x\n", id, drv_stat); |
913 | |
914 | /* We need a sensible error return here, which is tricky, and one |
915 | that won't cause people to do things like return a disk wrongly */ |
916 | *sk = ABORTED_COMMAND; |
917 | *asc = 0x00; |
918 | *ascq = 0x00; |
919 | |
920 | translate_done: |
921 | if (verbose) |
922 | printk(KERN_ERR "ata%u: translated ATA stat/err 0x%02x/%02x " |
923 | "to SCSI SK/ASC/ASCQ 0x%x/%02x/%02x\n", |
924 | id, drv_stat, drv_err, *sk, *asc, *ascq); |
925 | return; |
926 | } |
927 | |
928 | /* |
929 | * ata_gen_passthru_sense - Generate check condition sense block. |
930 | * @qc: Command that completed. |
931 | * |
932 | * This function is specific to the ATA descriptor format sense |
933 | * block specified for the ATA pass through commands. Regardless |
934 | * of whether the command errored or not, return a sense |
935 | * block. Copy all controller registers into the sense |
936 | * block. If there was no error, we get the request from an ATA |
937 | * passthrough command, so we use the following sense data: |
938 | * sk = RECOVERED ERROR |
939 | * asc,ascq = ATA PASS-THROUGH INFORMATION AVAILABLE |
940 | * |
941 | * |
942 | * LOCKING: |
943 | * None. |
944 | */ |
945 | static void ata_gen_passthru_sense(struct ata_queued_cmd *qc) |
946 | { |
947 | struct scsi_cmnd *cmd = qc->scsicmd; |
948 | struct ata_taskfile *tf = &qc->result_tf; |
949 | unsigned char *sb = cmd->sense_buffer; |
950 | unsigned char *desc = sb + 8; |
951 | int verbose = qc->ap->ops->error_handler == NULL; |
952 | |
953 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); |
954 | |
955 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
956 | |
957 | /* |
958 | * Use ata_to_sense_error() to map status register bits |
959 | * onto sense key, asc & ascq. |
960 | */ |
961 | if (qc->err_mask || |
962 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { |
963 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, |
964 | &sb[1], &sb[2], &sb[3], verbose); |
965 | sb[1] &= 0x0f; |
966 | } else { |
967 | sb[1] = RECOVERED_ERROR; |
968 | sb[2] = 0; |
969 | sb[3] = 0x1D; |
970 | } |
971 | |
972 | /* |
973 | * Sense data is current and format is descriptor. |
974 | */ |
975 | sb[0] = 0x72; |
976 | |
977 | desc[0] = 0x09; |
978 | |
979 | /* set length of additional sense data */ |
980 | sb[7] = 14; |
981 | desc[1] = 12; |
982 | |
983 | /* |
984 | * Copy registers into sense buffer. |
985 | */ |
986 | desc[2] = 0x00; |
987 | desc[3] = tf->feature; /* == error reg */ |
988 | desc[5] = tf->nsect; |
989 | desc[7] = tf->lbal; |
990 | desc[9] = tf->lbam; |
991 | desc[11] = tf->lbah; |
992 | desc[12] = tf->device; |
993 | desc[13] = tf->command; /* == status reg */ |
994 | |
995 | /* |
996 | * Fill in Extend bit, and the high order bytes |
997 | * if applicable. |
998 | */ |
999 | if (tf->flags & ATA_TFLAG_LBA48) { |
1000 | desc[2] |= 0x01; |
1001 | desc[4] = tf->hob_nsect; |
1002 | desc[6] = tf->hob_lbal; |
1003 | desc[8] = tf->hob_lbam; |
1004 | desc[10] = tf->hob_lbah; |
1005 | } |
1006 | } |
1007 | |
1008 | /** |
1009 | * ata_gen_ata_sense - generate a SCSI fixed sense block |
1010 | * @qc: Command that we are erroring out |
1011 | * |
1012 | * Generate sense block for a failed ATA command @qc. Descriptor |
1013 | * format is used to accommodate LBA48 block address. |
1014 | * |
1015 | * LOCKING: |
1016 | * None. |
1017 | */ |
1018 | static void ata_gen_ata_sense(struct ata_queued_cmd *qc) |
1019 | { |
1020 | struct ata_device *dev = qc->dev; |
1021 | struct scsi_cmnd *cmd = qc->scsicmd; |
1022 | struct ata_taskfile *tf = &qc->result_tf; |
1023 | unsigned char *sb = cmd->sense_buffer; |
1024 | unsigned char *desc = sb + 8; |
1025 | int verbose = qc->ap->ops->error_handler == NULL; |
1026 | u64 block; |
1027 | |
1028 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); |
1029 | |
1030 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
1031 | |
1032 | /* sense data is current and format is descriptor */ |
1033 | sb[0] = 0x72; |
1034 | |
1035 | /* Use ata_to_sense_error() to map status register bits |
1036 | * onto sense key, asc & ascq. |
1037 | */ |
1038 | if (qc->err_mask || |
1039 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { |
1040 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, |
1041 | &sb[1], &sb[2], &sb[3], verbose); |
1042 | sb[1] &= 0x0f; |
1043 | } |
1044 | |
1045 | block = ata_tf_read_block(&qc->result_tf, dev); |
1046 | |
1047 | /* information sense data descriptor */ |
1048 | sb[7] = 12; |
1049 | desc[0] = 0x00; |
1050 | desc[1] = 10; |
1051 | |
1052 | desc[2] |= 0x80; /* valid */ |
1053 | desc[6] = block >> 40; |
1054 | desc[7] = block >> 32; |
1055 | desc[8] = block >> 24; |
1056 | desc[9] = block >> 16; |
1057 | desc[10] = block >> 8; |
1058 | desc[11] = block; |
1059 | } |
1060 | |
1061 | static void ata_scsi_sdev_config(struct scsi_device *sdev) |
1062 | { |
1063 | sdev->use_10_for_rw = 1; |
1064 | sdev->use_10_for_ms = 1; |
1065 | sdev->no_report_opcodes = 1; |
1066 | sdev->no_write_same = 1; |
1067 | |
1068 | /* Schedule policy is determined by ->qc_defer() callback and |
1069 | * it needs to see every deferred qc. Set dev_blocked to 1 to |
1070 | * prevent SCSI midlayer from automatically deferring |
1071 | * requests. |
1072 | */ |
1073 | sdev->max_device_blocked = 1; |
1074 | } |
1075 | |
1076 | /** |
1077 | * atapi_drain_needed - Check whether data transfer may overflow |
1078 | * @rq: request to be checked |
1079 | * |
1080 | * ATAPI commands which transfer variable length data to host |
1081 | * might overflow due to application error or hardare bug. This |
1082 | * function checks whether overflow should be drained and ignored |
1083 | * for @request. |
1084 | * |
1085 | * LOCKING: |
1086 | * None. |
1087 | * |
1088 | * RETURNS: |
1089 | * 1 if ; otherwise, 0. |
1090 | */ |
1091 | static int atapi_drain_needed(struct request *rq) |
1092 | { |
1093 | if (likely(rq->cmd_type != REQ_TYPE_BLOCK_PC)) |
1094 | return 0; |
1095 | |
1096 | if (!blk_rq_bytes(rq) || (rq->cmd_flags & REQ_WRITE)) |
1097 | return 0; |
1098 | |
1099 | return atapi_cmd_type(rq->cmd[0]) == ATAPI_MISC; |
1100 | } |
1101 | |
1102 | static int ata_scsi_dev_config(struct scsi_device *sdev, |
1103 | struct ata_device *dev) |
1104 | { |
1105 | struct request_queue *q = sdev->request_queue; |
1106 | |
1107 | if (!ata_id_has_unload(dev->id)) |
1108 | dev->flags |= ATA_DFLAG_NO_UNLOAD; |
1109 | |
1110 | /* configure max sectors */ |
1111 | blk_queue_max_hw_sectors(q, dev->max_sectors); |
1112 | |
1113 | if (dev->class == ATA_DEV_ATAPI) { |
1114 | void *buf; |
1115 | |
1116 | sdev->sector_size = ATA_SECT_SIZE; |
1117 | |
1118 | /* set DMA padding */ |
1119 | blk_queue_update_dma_pad(q, ATA_DMA_PAD_SZ - 1); |
1120 | |
1121 | /* configure draining */ |
1122 | buf = kmalloc(ATAPI_MAX_DRAIN, q->bounce_gfp | GFP_KERNEL); |
1123 | if (!buf) { |
1124 | ata_dev_err(dev, "drain buffer allocation failed\n"); |
1125 | return -ENOMEM; |
1126 | } |
1127 | |
1128 | blk_queue_dma_drain(q, atapi_drain_needed, buf, ATAPI_MAX_DRAIN); |
1129 | } else { |
1130 | sdev->sector_size = ata_id_logical_sector_size(dev->id); |
1131 | sdev->manage_start_stop = 1; |
1132 | } |
1133 | |
1134 | /* |
1135 | * ata_pio_sectors() expects buffer for each sector to not cross |
1136 | * page boundary. Enforce it by requiring buffers to be sector |
1137 | * aligned, which works iff sector_size is not larger than |
1138 | * PAGE_SIZE. ATAPI devices also need the alignment as |
1139 | * IDENTIFY_PACKET is executed as ATA_PROT_PIO. |
1140 | */ |
1141 | if (sdev->sector_size > PAGE_SIZE) |
1142 | ata_dev_warn(dev, |
1143 | "sector_size=%u > PAGE_SIZE, PIO may malfunction\n", |
1144 | sdev->sector_size); |
1145 | |
1146 | blk_queue_update_dma_alignment(q, sdev->sector_size - 1); |
1147 | |
1148 | if (dev->flags & ATA_DFLAG_AN) |
1149 | set_bit(SDEV_EVT_MEDIA_CHANGE, sdev->supported_events); |
1150 | |
1151 | if (dev->flags & ATA_DFLAG_NCQ) { |
1152 | int depth; |
1153 | |
1154 | depth = min(sdev->host->can_queue, ata_id_queue_depth(dev->id)); |
1155 | depth = min(ATA_MAX_QUEUE - 1, depth); |
1156 | scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG, depth); |
1157 | } |
1158 | |
1159 | blk_queue_flush_queueable(q, false); |
1160 | |
1161 | dev->sdev = sdev; |
1162 | return 0; |
1163 | } |
1164 | |
1165 | /** |
1166 | * ata_scsi_slave_config - Set SCSI device attributes |
1167 | * @sdev: SCSI device to examine |
1168 | * |
1169 | * This is called before we actually start reading |
1170 | * and writing to the device, to configure certain |
1171 | * SCSI mid-layer behaviors. |
1172 | * |
1173 | * LOCKING: |
1174 | * Defined by SCSI layer. We don't really care. |
1175 | */ |
1176 | |
1177 | int ata_scsi_slave_config(struct scsi_device *sdev) |
1178 | { |
1179 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1180 | struct ata_device *dev = __ata_scsi_find_dev(ap, sdev); |
1181 | int rc = 0; |
1182 | |
1183 | ata_scsi_sdev_config(sdev); |
1184 | |
1185 | if (dev) |
1186 | rc = ata_scsi_dev_config(sdev, dev); |
1187 | |
1188 | return rc; |
1189 | } |
1190 | |
1191 | /** |
1192 | * ata_scsi_slave_destroy - SCSI device is about to be destroyed |
1193 | * @sdev: SCSI device to be destroyed |
1194 | * |
1195 | * @sdev is about to be destroyed for hot/warm unplugging. If |
1196 | * this unplugging was initiated by libata as indicated by NULL |
1197 | * dev->sdev, this function doesn't have to do anything. |
1198 | * Otherwise, SCSI layer initiated warm-unplug is in progress. |
1199 | * Clear dev->sdev, schedule the device for ATA detach and invoke |
1200 | * EH. |
1201 | * |
1202 | * LOCKING: |
1203 | * Defined by SCSI layer. We don't really care. |
1204 | */ |
1205 | void ata_scsi_slave_destroy(struct scsi_device *sdev) |
1206 | { |
1207 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1208 | struct request_queue *q = sdev->request_queue; |
1209 | unsigned long flags; |
1210 | struct ata_device *dev; |
1211 | |
1212 | if (!ap->ops->error_handler) |
1213 | return; |
1214 | |
1215 | spin_lock_irqsave(ap->lock, flags); |
1216 | dev = __ata_scsi_find_dev(ap, sdev); |
1217 | if (dev && dev->sdev) { |
1218 | /* SCSI device already in CANCEL state, no need to offline it */ |
1219 | dev->sdev = NULL; |
1220 | dev->flags |= ATA_DFLAG_DETACH; |
1221 | ata_port_schedule_eh(ap); |
1222 | } |
1223 | spin_unlock_irqrestore(ap->lock, flags); |
1224 | |
1225 | kfree(q->dma_drain_buffer); |
1226 | q->dma_drain_buffer = NULL; |
1227 | q->dma_drain_size = 0; |
1228 | } |
1229 | |
1230 | /** |
1231 | * __ata_change_queue_depth - helper for ata_scsi_change_queue_depth |
1232 | * @ap: ATA port to which the device change the queue depth |
1233 | * @sdev: SCSI device to configure queue depth for |
1234 | * @queue_depth: new queue depth |
1235 | * @reason: calling context |
1236 | * |
1237 | * libsas and libata have different approaches for associating a sdev to |
1238 | * its ata_port. |
1239 | * |
1240 | */ |
1241 | int __ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, |
1242 | int queue_depth, int reason) |
1243 | { |
1244 | struct ata_device *dev; |
1245 | unsigned long flags; |
1246 | |
1247 | if (reason != SCSI_QDEPTH_DEFAULT) |
1248 | return -EOPNOTSUPP; |
1249 | |
1250 | if (queue_depth < 1 || queue_depth == sdev->queue_depth) |
1251 | return sdev->queue_depth; |
1252 | |
1253 | dev = ata_scsi_find_dev(ap, sdev); |
1254 | if (!dev || !ata_dev_enabled(dev)) |
1255 | return sdev->queue_depth; |
1256 | |
1257 | /* NCQ enabled? */ |
1258 | spin_lock_irqsave(ap->lock, flags); |
1259 | dev->flags &= ~ATA_DFLAG_NCQ_OFF; |
1260 | if (queue_depth == 1 || !ata_ncq_enabled(dev)) { |
1261 | dev->flags |= ATA_DFLAG_NCQ_OFF; |
1262 | queue_depth = 1; |
1263 | } |
1264 | spin_unlock_irqrestore(ap->lock, flags); |
1265 | |
1266 | /* limit and apply queue depth */ |
1267 | queue_depth = min(queue_depth, sdev->host->can_queue); |
1268 | queue_depth = min(queue_depth, ata_id_queue_depth(dev->id)); |
1269 | queue_depth = min(queue_depth, ATA_MAX_QUEUE - 1); |
1270 | |
1271 | if (sdev->queue_depth == queue_depth) |
1272 | return -EINVAL; |
1273 | |
1274 | scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG, queue_depth); |
1275 | return queue_depth; |
1276 | } |
1277 | |
1278 | /** |
1279 | * ata_scsi_change_queue_depth - SCSI callback for queue depth config |
1280 | * @sdev: SCSI device to configure queue depth for |
1281 | * @queue_depth: new queue depth |
1282 | * @reason: calling context |
1283 | * |
1284 | * This is libata standard hostt->change_queue_depth callback. |
1285 | * SCSI will call into this callback when user tries to set queue |
1286 | * depth via sysfs. |
1287 | * |
1288 | * LOCKING: |
1289 | * SCSI layer (we don't care) |
1290 | * |
1291 | * RETURNS: |
1292 | * Newly configured queue depth. |
1293 | */ |
1294 | int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth, |
1295 | int reason) |
1296 | { |
1297 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1298 | |
1299 | return __ata_change_queue_depth(ap, sdev, queue_depth, reason); |
1300 | } |
1301 | |
1302 | /** |
1303 | * ata_scsi_start_stop_xlat - Translate SCSI START STOP UNIT command |
1304 | * @qc: Storage for translated ATA taskfile |
1305 | * |
1306 | * Sets up an ATA taskfile to issue STANDBY (to stop) or READ VERIFY |
1307 | * (to start). Perhaps these commands should be preceded by |
1308 | * CHECK POWER MODE to see what power mode the device is already in. |
1309 | * [See SAT revision 5 at www.t10.org] |
1310 | * |
1311 | * LOCKING: |
1312 | * spin_lock_irqsave(host lock) |
1313 | * |
1314 | * RETURNS: |
1315 | * Zero on success, non-zero on error. |
1316 | */ |
1317 | static unsigned int ata_scsi_start_stop_xlat(struct ata_queued_cmd *qc) |
1318 | { |
1319 | struct scsi_cmnd *scmd = qc->scsicmd; |
1320 | struct ata_taskfile *tf = &qc->tf; |
1321 | const u8 *cdb = scmd->cmnd; |
1322 | |
1323 | if (scmd->cmd_len < 5) |
1324 | goto invalid_fld; |
1325 | |
1326 | tf->flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; |
1327 | tf->protocol = ATA_PROT_NODATA; |
1328 | if (cdb[1] & 0x1) { |
1329 | ; /* ignore IMMED bit, violates sat-r05 */ |
1330 | } |
1331 | if (cdb[4] & 0x2) |
1332 | goto invalid_fld; /* LOEJ bit set not supported */ |
1333 | if (((cdb[4] >> 4) & 0xf) != 0) |
1334 | goto invalid_fld; /* power conditions not supported */ |
1335 | |
1336 | if (cdb[4] & 0x1) { |
1337 | tf->nsect = 1; /* 1 sector, lba=0 */ |
1338 | |
1339 | if (qc->dev->flags & ATA_DFLAG_LBA) { |
1340 | tf->flags |= ATA_TFLAG_LBA; |
1341 | |
1342 | tf->lbah = 0x0; |
1343 | tf->lbam = 0x0; |
1344 | tf->lbal = 0x0; |
1345 | tf->device |= ATA_LBA; |
1346 | } else { |
1347 | /* CHS */ |
1348 | tf->lbal = 0x1; /* sect */ |
1349 | tf->lbam = 0x0; /* cyl low */ |
1350 | tf->lbah = 0x0; /* cyl high */ |
1351 | } |
1352 | |
1353 | tf->command = ATA_CMD_VERIFY; /* READ VERIFY */ |
1354 | } else { |
1355 | /* Some odd clown BIOSen issue spindown on power off (ACPI S4 |
1356 | * or S5) causing some drives to spin up and down again. |
1357 | */ |
1358 | if ((qc->ap->flags & ATA_FLAG_NO_POWEROFF_SPINDOWN) && |
1359 | system_state == SYSTEM_POWER_OFF) |
1360 | goto skip; |
1361 | |
1362 | if ((qc->ap->flags & ATA_FLAG_NO_HIBERNATE_SPINDOWN) && |
1363 | system_entering_hibernation()) |
1364 | goto skip; |
1365 | |
1366 | /* Issue ATA STANDBY IMMEDIATE command */ |
1367 | tf->command = ATA_CMD_STANDBYNOW1; |
1368 | } |
1369 | |
1370 | /* |
1371 | * Standby and Idle condition timers could be implemented but that |
1372 | * would require libata to implement the Power condition mode page |
1373 | * and allow the user to change it. Changing mode pages requires |
1374 | * MODE SELECT to be implemented. |
1375 | */ |
1376 | |
1377 | return 0; |
1378 | |
1379 | invalid_fld: |
1380 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1381 | /* "Invalid field in cbd" */ |
1382 | return 1; |
1383 | skip: |
1384 | scmd->result = SAM_STAT_GOOD; |
1385 | return 1; |
1386 | } |
1387 | |
1388 | |
1389 | /** |
1390 | * ata_scsi_flush_xlat - Translate SCSI SYNCHRONIZE CACHE command |
1391 | * @qc: Storage for translated ATA taskfile |
1392 | * |
1393 | * Sets up an ATA taskfile to issue FLUSH CACHE or |
1394 | * FLUSH CACHE EXT. |
1395 | * |
1396 | * LOCKING: |
1397 | * spin_lock_irqsave(host lock) |
1398 | * |
1399 | * RETURNS: |
1400 | * Zero on success, non-zero on error. |
1401 | */ |
1402 | static unsigned int ata_scsi_flush_xlat(struct ata_queued_cmd *qc) |
1403 | { |
1404 | struct ata_taskfile *tf = &qc->tf; |
1405 | |
1406 | tf->flags |= ATA_TFLAG_DEVICE; |
1407 | tf->protocol = ATA_PROT_NODATA; |
1408 | |
1409 | if (qc->dev->flags & ATA_DFLAG_FLUSH_EXT) |
1410 | tf->command = ATA_CMD_FLUSH_EXT; |
1411 | else |
1412 | tf->command = ATA_CMD_FLUSH; |
1413 | |
1414 | /* flush is critical for IO integrity, consider it an IO command */ |
1415 | qc->flags |= ATA_QCFLAG_IO; |
1416 | |
1417 | return 0; |
1418 | } |
1419 | |
1420 | /** |
1421 | * scsi_6_lba_len - Get LBA and transfer length |
1422 | * @cdb: SCSI command to translate |
1423 | * |
1424 | * Calculate LBA and transfer length for 6-byte commands. |
1425 | * |
1426 | * RETURNS: |
1427 | * @plba: the LBA |
1428 | * @plen: the transfer length |
1429 | */ |
1430 | static void scsi_6_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1431 | { |
1432 | u64 lba = 0; |
1433 | u32 len; |
1434 | |
1435 | VPRINTK("six-byte command\n"); |
1436 | |
1437 | lba |= ((u64)(cdb[1] & 0x1f)) << 16; |
1438 | lba |= ((u64)cdb[2]) << 8; |
1439 | lba |= ((u64)cdb[3]); |
1440 | |
1441 | len = cdb[4]; |
1442 | |
1443 | *plba = lba; |
1444 | *plen = len; |
1445 | } |
1446 | |
1447 | /** |
1448 | * scsi_10_lba_len - Get LBA and transfer length |
1449 | * @cdb: SCSI command to translate |
1450 | * |
1451 | * Calculate LBA and transfer length for 10-byte commands. |
1452 | * |
1453 | * RETURNS: |
1454 | * @plba: the LBA |
1455 | * @plen: the transfer length |
1456 | */ |
1457 | static void scsi_10_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1458 | { |
1459 | u64 lba = 0; |
1460 | u32 len = 0; |
1461 | |
1462 | VPRINTK("ten-byte command\n"); |
1463 | |
1464 | lba |= ((u64)cdb[2]) << 24; |
1465 | lba |= ((u64)cdb[3]) << 16; |
1466 | lba |= ((u64)cdb[4]) << 8; |
1467 | lba |= ((u64)cdb[5]); |
1468 | |
1469 | len |= ((u32)cdb[7]) << 8; |
1470 | len |= ((u32)cdb[8]); |
1471 | |
1472 | *plba = lba; |
1473 | *plen = len; |
1474 | } |
1475 | |
1476 | /** |
1477 | * scsi_16_lba_len - Get LBA and transfer length |
1478 | * @cdb: SCSI command to translate |
1479 | * |
1480 | * Calculate LBA and transfer length for 16-byte commands. |
1481 | * |
1482 | * RETURNS: |
1483 | * @plba: the LBA |
1484 | * @plen: the transfer length |
1485 | */ |
1486 | static void scsi_16_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1487 | { |
1488 | u64 lba = 0; |
1489 | u32 len = 0; |
1490 | |
1491 | VPRINTK("sixteen-byte command\n"); |
1492 | |
1493 | lba |= ((u64)cdb[2]) << 56; |
1494 | lba |= ((u64)cdb[3]) << 48; |
1495 | lba |= ((u64)cdb[4]) << 40; |
1496 | lba |= ((u64)cdb[5]) << 32; |
1497 | lba |= ((u64)cdb[6]) << 24; |
1498 | lba |= ((u64)cdb[7]) << 16; |
1499 | lba |= ((u64)cdb[8]) << 8; |
1500 | lba |= ((u64)cdb[9]); |
1501 | |
1502 | len |= ((u32)cdb[10]) << 24; |
1503 | len |= ((u32)cdb[11]) << 16; |
1504 | len |= ((u32)cdb[12]) << 8; |
1505 | len |= ((u32)cdb[13]); |
1506 | |
1507 | *plba = lba; |
1508 | *plen = len; |
1509 | } |
1510 | |
1511 | /** |
1512 | * ata_scsi_verify_xlat - Translate SCSI VERIFY command into an ATA one |
1513 | * @qc: Storage for translated ATA taskfile |
1514 | * |
1515 | * Converts SCSI VERIFY command to an ATA READ VERIFY command. |
1516 | * |
1517 | * LOCKING: |
1518 | * spin_lock_irqsave(host lock) |
1519 | * |
1520 | * RETURNS: |
1521 | * Zero on success, non-zero on error. |
1522 | */ |
1523 | static unsigned int ata_scsi_verify_xlat(struct ata_queued_cmd *qc) |
1524 | { |
1525 | struct scsi_cmnd *scmd = qc->scsicmd; |
1526 | struct ata_taskfile *tf = &qc->tf; |
1527 | struct ata_device *dev = qc->dev; |
1528 | u64 dev_sectors = qc->dev->n_sectors; |
1529 | const u8 *cdb = scmd->cmnd; |
1530 | u64 block; |
1531 | u32 n_block; |
1532 | |
1533 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
1534 | tf->protocol = ATA_PROT_NODATA; |
1535 | |
1536 | if (cdb[0] == VERIFY) { |
1537 | if (scmd->cmd_len < 10) |
1538 | goto invalid_fld; |
1539 | scsi_10_lba_len(cdb, &block, &n_block); |
1540 | } else if (cdb[0] == VERIFY_16) { |
1541 | if (scmd->cmd_len < 16) |
1542 | goto invalid_fld; |
1543 | scsi_16_lba_len(cdb, &block, &n_block); |
1544 | } else |
1545 | goto invalid_fld; |
1546 | |
1547 | if (!n_block) |
1548 | goto nothing_to_do; |
1549 | if (block >= dev_sectors) |
1550 | goto out_of_range; |
1551 | if ((block + n_block) > dev_sectors) |
1552 | goto out_of_range; |
1553 | |
1554 | if (dev->flags & ATA_DFLAG_LBA) { |
1555 | tf->flags |= ATA_TFLAG_LBA; |
1556 | |
1557 | if (lba_28_ok(block, n_block)) { |
1558 | /* use LBA28 */ |
1559 | tf->command = ATA_CMD_VERIFY; |
1560 | tf->device |= (block >> 24) & 0xf; |
1561 | } else if (lba_48_ok(block, n_block)) { |
1562 | if (!(dev->flags & ATA_DFLAG_LBA48)) |
1563 | goto out_of_range; |
1564 | |
1565 | /* use LBA48 */ |
1566 | tf->flags |= ATA_TFLAG_LBA48; |
1567 | tf->command = ATA_CMD_VERIFY_EXT; |
1568 | |
1569 | tf->hob_nsect = (n_block >> 8) & 0xff; |
1570 | |
1571 | tf->hob_lbah = (block >> 40) & 0xff; |
1572 | tf->hob_lbam = (block >> 32) & 0xff; |
1573 | tf->hob_lbal = (block >> 24) & 0xff; |
1574 | } else |
1575 | /* request too large even for LBA48 */ |
1576 | goto out_of_range; |
1577 | |
1578 | tf->nsect = n_block & 0xff; |
1579 | |
1580 | tf->lbah = (block >> 16) & 0xff; |
1581 | tf->lbam = (block >> 8) & 0xff; |
1582 | tf->lbal = block & 0xff; |
1583 | |
1584 | tf->device |= ATA_LBA; |
1585 | } else { |
1586 | /* CHS */ |
1587 | u32 sect, head, cyl, track; |
1588 | |
1589 | if (!lba_28_ok(block, n_block)) |
1590 | goto out_of_range; |
1591 | |
1592 | /* Convert LBA to CHS */ |
1593 | track = (u32)block / dev->sectors; |
1594 | cyl = track / dev->heads; |
1595 | head = track % dev->heads; |
1596 | sect = (u32)block % dev->sectors + 1; |
1597 | |
1598 | DPRINTK("block %u track %u cyl %u head %u sect %u\n", |
1599 | (u32)block, track, cyl, head, sect); |
1600 | |
1601 | /* Check whether the converted CHS can fit. |
1602 | Cylinder: 0-65535 |
1603 | Head: 0-15 |
1604 | Sector: 1-255*/ |
1605 | if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect)) |
1606 | goto out_of_range; |
1607 | |
1608 | tf->command = ATA_CMD_VERIFY; |
1609 | tf->nsect = n_block & 0xff; /* Sector count 0 means 256 sectors */ |
1610 | tf->lbal = sect; |
1611 | tf->lbam = cyl; |
1612 | tf->lbah = cyl >> 8; |
1613 | tf->device |= head; |
1614 | } |
1615 | |
1616 | return 0; |
1617 | |
1618 | invalid_fld: |
1619 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1620 | /* "Invalid field in cbd" */ |
1621 | return 1; |
1622 | |
1623 | out_of_range: |
1624 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); |
1625 | /* "Logical Block Address out of range" */ |
1626 | return 1; |
1627 | |
1628 | nothing_to_do: |
1629 | scmd->result = SAM_STAT_GOOD; |
1630 | return 1; |
1631 | } |
1632 | |
1633 | /** |
1634 | * ata_scsi_rw_xlat - Translate SCSI r/w command into an ATA one |
1635 | * @qc: Storage for translated ATA taskfile |
1636 | * |
1637 | * Converts any of six SCSI read/write commands into the |
1638 | * ATA counterpart, including starting sector (LBA), |
1639 | * sector count, and taking into account the device's LBA48 |
1640 | * support. |
1641 | * |
1642 | * Commands %READ_6, %READ_10, %READ_16, %WRITE_6, %WRITE_10, and |
1643 | * %WRITE_16 are currently supported. |
1644 | * |
1645 | * LOCKING: |
1646 | * spin_lock_irqsave(host lock) |
1647 | * |
1648 | * RETURNS: |
1649 | * Zero on success, non-zero on error. |
1650 | */ |
1651 | static unsigned int ata_scsi_rw_xlat(struct ata_queued_cmd *qc) |
1652 | { |
1653 | struct scsi_cmnd *scmd = qc->scsicmd; |
1654 | const u8 *cdb = scmd->cmnd; |
1655 | unsigned int tf_flags = 0; |
1656 | u64 block; |
1657 | u32 n_block; |
1658 | int rc; |
1659 | |
1660 | if (cdb[0] == WRITE_10 || cdb[0] == WRITE_6 || cdb[0] == WRITE_16) |
1661 | tf_flags |= ATA_TFLAG_WRITE; |
1662 | |
1663 | /* Calculate the SCSI LBA, transfer length and FUA. */ |
1664 | switch (cdb[0]) { |
1665 | case READ_10: |
1666 | case WRITE_10: |
1667 | if (unlikely(scmd->cmd_len < 10)) |
1668 | goto invalid_fld; |
1669 | scsi_10_lba_len(cdb, &block, &n_block); |
1670 | if (cdb[1] & (1 << 3)) |
1671 | tf_flags |= ATA_TFLAG_FUA; |
1672 | break; |
1673 | case READ_6: |
1674 | case WRITE_6: |
1675 | if (unlikely(scmd->cmd_len < 6)) |
1676 | goto invalid_fld; |
1677 | scsi_6_lba_len(cdb, &block, &n_block); |
1678 | |
1679 | /* for 6-byte r/w commands, transfer length 0 |
1680 | * means 256 blocks of data, not 0 block. |
1681 | */ |
1682 | if (!n_block) |
1683 | n_block = 256; |
1684 | break; |
1685 | case READ_16: |
1686 | case WRITE_16: |
1687 | if (unlikely(scmd->cmd_len < 16)) |
1688 | goto invalid_fld; |
1689 | scsi_16_lba_len(cdb, &block, &n_block); |
1690 | if (cdb[1] & (1 << 3)) |
1691 | tf_flags |= ATA_TFLAG_FUA; |
1692 | break; |
1693 | default: |
1694 | DPRINTK("no-byte command\n"); |
1695 | goto invalid_fld; |
1696 | } |
1697 | |
1698 | /* Check and compose ATA command */ |
1699 | if (!n_block) |
1700 | /* For 10-byte and 16-byte SCSI R/W commands, transfer |
1701 | * length 0 means transfer 0 block of data. |
1702 | * However, for ATA R/W commands, sector count 0 means |
1703 | * 256 or 65536 sectors, not 0 sectors as in SCSI. |
1704 | * |
1705 | * WARNING: one or two older ATA drives treat 0 as 0... |
1706 | */ |
1707 | goto nothing_to_do; |
1708 | |
1709 | qc->flags |= ATA_QCFLAG_IO; |
1710 | qc->nbytes = n_block * scmd->device->sector_size; |
1711 | |
1712 | rc = ata_build_rw_tf(&qc->tf, qc->dev, block, n_block, tf_flags, |
1713 | qc->tag); |
1714 | if (likely(rc == 0)) |
1715 | return 0; |
1716 | |
1717 | if (rc == -ERANGE) |
1718 | goto out_of_range; |
1719 | /* treat all other errors as -EINVAL, fall through */ |
1720 | invalid_fld: |
1721 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1722 | /* "Invalid field in cbd" */ |
1723 | return 1; |
1724 | |
1725 | out_of_range: |
1726 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); |
1727 | /* "Logical Block Address out of range" */ |
1728 | return 1; |
1729 | |
1730 | nothing_to_do: |
1731 | scmd->result = SAM_STAT_GOOD; |
1732 | return 1; |
1733 | } |
1734 | |
1735 | static void ata_scsi_qc_complete(struct ata_queued_cmd *qc) |
1736 | { |
1737 | struct ata_port *ap = qc->ap; |
1738 | struct scsi_cmnd *cmd = qc->scsicmd; |
1739 | u8 *cdb = cmd->cmnd; |
1740 | int need_sense = (qc->err_mask != 0); |
1741 | |
1742 | /* For ATA pass thru (SAT) commands, generate a sense block if |
1743 | * user mandated it or if there's an error. Note that if we |
1744 | * generate because the user forced us to [CK_COND =1], a check |
1745 | * condition is generated and the ATA register values are returned |
1746 | * whether the command completed successfully or not. If there |
1747 | * was no error, we use the following sense data: |
1748 | * sk = RECOVERED ERROR |
1749 | * asc,ascq = ATA PASS-THROUGH INFORMATION AVAILABLE |
1750 | */ |
1751 | if (((cdb[0] == ATA_16) || (cdb[0] == ATA_12)) && |
1752 | ((cdb[2] & 0x20) || need_sense)) { |
1753 | ata_gen_passthru_sense(qc); |
1754 | } else { |
1755 | if (!need_sense) { |
1756 | cmd->result = SAM_STAT_GOOD; |
1757 | } else { |
1758 | /* TODO: decide which descriptor format to use |
1759 | * for 48b LBA devices and call that here |
1760 | * instead of the fixed desc, which is only |
1761 | * good for smaller LBA (and maybe CHS?) |
1762 | * devices. |
1763 | */ |
1764 | ata_gen_ata_sense(qc); |
1765 | } |
1766 | } |
1767 | |
1768 | if (need_sense && !ap->ops->error_handler) |
1769 | ata_dump_status(ap->print_id, &qc->result_tf); |
1770 | |
1771 | qc->scsidone(cmd); |
1772 | |
1773 | ata_qc_free(qc); |
1774 | } |
1775 | |
1776 | /** |
1777 | * ata_scsi_translate - Translate then issue SCSI command to ATA device |
1778 | * @dev: ATA device to which the command is addressed |
1779 | * @cmd: SCSI command to execute |
1780 | * @xlat_func: Actor which translates @cmd to an ATA taskfile |
1781 | * |
1782 | * Our ->queuecommand() function has decided that the SCSI |
1783 | * command issued can be directly translated into an ATA |
1784 | * command, rather than handled internally. |
1785 | * |
1786 | * This function sets up an ata_queued_cmd structure for the |
1787 | * SCSI command, and sends that ata_queued_cmd to the hardware. |
1788 | * |
1789 | * The xlat_func argument (actor) returns 0 if ready to execute |
1790 | * ATA command, else 1 to finish translation. If 1 is returned |
1791 | * then cmd->result (and possibly cmd->sense_buffer) are assumed |
1792 | * to be set reflecting an error condition or clean (early) |
1793 | * termination. |
1794 | * |
1795 | * LOCKING: |
1796 | * spin_lock_irqsave(host lock) |
1797 | * |
1798 | * RETURNS: |
1799 | * 0 on success, SCSI_ML_QUEUE_DEVICE_BUSY if the command |
1800 | * needs to be deferred. |
1801 | */ |
1802 | static int ata_scsi_translate(struct ata_device *dev, struct scsi_cmnd *cmd, |
1803 | ata_xlat_func_t xlat_func) |
1804 | { |
1805 | struct ata_port *ap = dev->link->ap; |
1806 | struct ata_queued_cmd *qc; |
1807 | int rc; |
1808 | |
1809 | VPRINTK("ENTER\n"); |
1810 | |
1811 | qc = ata_scsi_qc_new(dev, cmd); |
1812 | if (!qc) |
1813 | goto err_mem; |
1814 | |
1815 | /* data is present; dma-map it */ |
1816 | if (cmd->sc_data_direction == DMA_FROM_DEVICE || |
1817 | cmd->sc_data_direction == DMA_TO_DEVICE) { |
1818 | if (unlikely(scsi_bufflen(cmd) < 1)) { |
1819 | ata_dev_warn(dev, "WARNING: zero len r/w req\n"); |
1820 | goto err_did; |
1821 | } |
1822 | |
1823 | ata_sg_init(qc, scsi_sglist(cmd), scsi_sg_count(cmd)); |
1824 | |
1825 | qc->dma_dir = cmd->sc_data_direction; |
1826 | } |
1827 | |
1828 | qc->complete_fn = ata_scsi_qc_complete; |
1829 | |
1830 | if (xlat_func(qc)) |
1831 | goto early_finish; |
1832 | |
1833 | if (ap->ops->qc_defer) { |
1834 | if ((rc = ap->ops->qc_defer(qc))) |
1835 | goto defer; |
1836 | } |
1837 | |
1838 | /* select device, send command to hardware */ |
1839 | ata_qc_issue(qc); |
1840 | |
1841 | VPRINTK("EXIT\n"); |
1842 | return 0; |
1843 | |
1844 | early_finish: |
1845 | ata_qc_free(qc); |
1846 | cmd->scsi_done(cmd); |
1847 | DPRINTK("EXIT - early finish (good or error)\n"); |
1848 | return 0; |
1849 | |
1850 | err_did: |
1851 | ata_qc_free(qc); |
1852 | cmd->result = (DID_ERROR << 16); |
1853 | cmd->scsi_done(cmd); |
1854 | err_mem: |
1855 | DPRINTK("EXIT - internal\n"); |
1856 | return 0; |
1857 | |
1858 | defer: |
1859 | ata_qc_free(qc); |
1860 | DPRINTK("EXIT - defer\n"); |
1861 | if (rc == ATA_DEFER_LINK) |
1862 | return SCSI_MLQUEUE_DEVICE_BUSY; |
1863 | else |
1864 | return SCSI_MLQUEUE_HOST_BUSY; |
1865 | } |
1866 | |
1867 | /** |
1868 | * ata_scsi_rbuf_get - Map response buffer. |
1869 | * @cmd: SCSI command containing buffer to be mapped. |
1870 | * @flags: unsigned long variable to store irq enable status |
1871 | * @copy_in: copy in from user buffer |
1872 | * |
1873 | * Prepare buffer for simulated SCSI commands. |
1874 | * |
1875 | * LOCKING: |
1876 | * spin_lock_irqsave(ata_scsi_rbuf_lock) on success |
1877 | * |
1878 | * RETURNS: |
1879 | * Pointer to response buffer. |
1880 | */ |
1881 | static void *ata_scsi_rbuf_get(struct scsi_cmnd *cmd, bool copy_in, |
1882 | unsigned long *flags) |
1883 | { |
1884 | spin_lock_irqsave(&ata_scsi_rbuf_lock, *flags); |
1885 | |
1886 | memset(ata_scsi_rbuf, 0, ATA_SCSI_RBUF_SIZE); |
1887 | if (copy_in) |
1888 | sg_copy_to_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), |
1889 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); |
1890 | return ata_scsi_rbuf; |
1891 | } |
1892 | |
1893 | /** |
1894 | * ata_scsi_rbuf_put - Unmap response buffer. |
1895 | * @cmd: SCSI command containing buffer to be unmapped. |
1896 | * @copy_out: copy out result |
1897 | * @flags: @flags passed to ata_scsi_rbuf_get() |
1898 | * |
1899 | * Returns rbuf buffer. The result is copied to @cmd's buffer if |
1900 | * @copy_back is true. |
1901 | * |
1902 | * LOCKING: |
1903 | * Unlocks ata_scsi_rbuf_lock. |
1904 | */ |
1905 | static inline void ata_scsi_rbuf_put(struct scsi_cmnd *cmd, bool copy_out, |
1906 | unsigned long *flags) |
1907 | { |
1908 | if (copy_out) |
1909 | sg_copy_from_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), |
1910 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); |
1911 | spin_unlock_irqrestore(&ata_scsi_rbuf_lock, *flags); |
1912 | } |
1913 | |
1914 | /** |
1915 | * ata_scsi_rbuf_fill - wrapper for SCSI command simulators |
1916 | * @args: device IDENTIFY data / SCSI command of interest. |
1917 | * @actor: Callback hook for desired SCSI command simulator |
1918 | * |
1919 | * Takes care of the hard work of simulating a SCSI command... |
1920 | * Mapping the response buffer, calling the command's handler, |
1921 | * and handling the handler's return value. This return value |
1922 | * indicates whether the handler wishes the SCSI command to be |
1923 | * completed successfully (0), or not (in which case cmd->result |
1924 | * and sense buffer are assumed to be set). |
1925 | * |
1926 | * LOCKING: |
1927 | * spin_lock_irqsave(host lock) |
1928 | */ |
1929 | static void ata_scsi_rbuf_fill(struct ata_scsi_args *args, |
1930 | unsigned int (*actor)(struct ata_scsi_args *args, u8 *rbuf)) |
1931 | { |
1932 | u8 *rbuf; |
1933 | unsigned int rc; |
1934 | struct scsi_cmnd *cmd = args->cmd; |
1935 | unsigned long flags; |
1936 | |
1937 | rbuf = ata_scsi_rbuf_get(cmd, false, &flags); |
1938 | rc = actor(args, rbuf); |
1939 | ata_scsi_rbuf_put(cmd, rc == 0, &flags); |
1940 | |
1941 | if (rc == 0) |
1942 | cmd->result = SAM_STAT_GOOD; |
1943 | args->done(cmd); |
1944 | } |
1945 | |
1946 | /** |
1947 | * ata_scsiop_inq_std - Simulate INQUIRY command |
1948 | * @args: device IDENTIFY data / SCSI command of interest. |
1949 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
1950 | * |
1951 | * Returns standard device identification data associated |
1952 | * with non-VPD INQUIRY command output. |
1953 | * |
1954 | * LOCKING: |
1955 | * spin_lock_irqsave(host lock) |
1956 | */ |
1957 | static unsigned int ata_scsiop_inq_std(struct ata_scsi_args *args, u8 *rbuf) |
1958 | { |
1959 | const u8 versions[] = { |
1960 | 0x60, /* SAM-3 (no version claimed) */ |
1961 | |
1962 | 0x03, |
1963 | 0x20, /* SBC-2 (no version claimed) */ |
1964 | |
1965 | 0x02, |
1966 | 0x60 /* SPC-3 (no version claimed) */ |
1967 | }; |
1968 | u8 hdr[] = { |
1969 | TYPE_DISK, |
1970 | 0, |
1971 | 0x5, /* claim SPC-3 version compatibility */ |
1972 | 2, |
1973 | 95 - 4 |
1974 | }; |
1975 | |
1976 | VPRINTK("ENTER\n"); |
1977 | |
1978 | /* set scsi removeable (RMB) bit per ata bit */ |
1979 | if (ata_id_removeable(args->id)) |
1980 | hdr[1] |= (1 << 7); |
1981 | |
1982 | memcpy(rbuf, hdr, sizeof(hdr)); |
1983 | memcpy(&rbuf[8], "ATA ", 8); |
1984 | ata_id_string(args->id, &rbuf[16], ATA_ID_PROD, 16); |
1985 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV, 4); |
1986 | |
1987 | if (rbuf[32] == 0 || rbuf[32] == ' ') |
1988 | memcpy(&rbuf[32], "n/a ", 4); |
1989 | |
1990 | memcpy(rbuf + 59, versions, sizeof(versions)); |
1991 | |
1992 | return 0; |
1993 | } |
1994 | |
1995 | /** |
1996 | * ata_scsiop_inq_00 - Simulate INQUIRY VPD page 0, list of pages |
1997 | * @args: device IDENTIFY data / SCSI command of interest. |
1998 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
1999 | * |
2000 | * Returns list of inquiry VPD pages available. |
2001 | * |
2002 | * LOCKING: |
2003 | * spin_lock_irqsave(host lock) |
2004 | */ |
2005 | static unsigned int ata_scsiop_inq_00(struct ata_scsi_args *args, u8 *rbuf) |
2006 | { |
2007 | const u8 pages[] = { |
2008 | 0x00, /* page 0x00, this page */ |
2009 | 0x80, /* page 0x80, unit serial no page */ |
2010 | 0x83, /* page 0x83, device ident page */ |
2011 | 0x89, /* page 0x89, ata info page */ |
2012 | 0xb0, /* page 0xb0, block limits page */ |
2013 | 0xb1, /* page 0xb1, block device characteristics page */ |
2014 | 0xb2, /* page 0xb2, thin provisioning page */ |
2015 | }; |
2016 | |
2017 | rbuf[3] = sizeof(pages); /* number of supported VPD pages */ |
2018 | memcpy(rbuf + 4, pages, sizeof(pages)); |
2019 | return 0; |
2020 | } |
2021 | |
2022 | /** |
2023 | * ata_scsiop_inq_80 - Simulate INQUIRY VPD page 80, device serial number |
2024 | * @args: device IDENTIFY data / SCSI command of interest. |
2025 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2026 | * |
2027 | * Returns ATA device serial number. |
2028 | * |
2029 | * LOCKING: |
2030 | * spin_lock_irqsave(host lock) |
2031 | */ |
2032 | static unsigned int ata_scsiop_inq_80(struct ata_scsi_args *args, u8 *rbuf) |
2033 | { |
2034 | const u8 hdr[] = { |
2035 | 0, |
2036 | 0x80, /* this page code */ |
2037 | 0, |
2038 | ATA_ID_SERNO_LEN, /* page len */ |
2039 | }; |
2040 | |
2041 | memcpy(rbuf, hdr, sizeof(hdr)); |
2042 | ata_id_string(args->id, (unsigned char *) &rbuf[4], |
2043 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
2044 | return 0; |
2045 | } |
2046 | |
2047 | /** |
2048 | * ata_scsiop_inq_83 - Simulate INQUIRY VPD page 83, device identity |
2049 | * @args: device IDENTIFY data / SCSI command of interest. |
2050 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2051 | * |
2052 | * Yields two logical unit device identification designators: |
2053 | * - vendor specific ASCII containing the ATA serial number |
2054 | * - SAT defined "t10 vendor id based" containing ASCII vendor |
2055 | * name ("ATA "), model and serial numbers. |
2056 | * |
2057 | * LOCKING: |
2058 | * spin_lock_irqsave(host lock) |
2059 | */ |
2060 | static unsigned int ata_scsiop_inq_83(struct ata_scsi_args *args, u8 *rbuf) |
2061 | { |
2062 | const int sat_model_serial_desc_len = 68; |
2063 | int num; |
2064 | |
2065 | rbuf[1] = 0x83; /* this page code */ |
2066 | num = 4; |
2067 | |
2068 | /* piv=0, assoc=lu, code_set=ACSII, designator=vendor */ |
2069 | rbuf[num + 0] = 2; |
2070 | rbuf[num + 3] = ATA_ID_SERNO_LEN; |
2071 | num += 4; |
2072 | ata_id_string(args->id, (unsigned char *) rbuf + num, |
2073 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
2074 | num += ATA_ID_SERNO_LEN; |
2075 | |
2076 | /* SAT defined lu model and serial numbers descriptor */ |
2077 | /* piv=0, assoc=lu, code_set=ACSII, designator=t10 vendor id */ |
2078 | rbuf[num + 0] = 2; |
2079 | rbuf[num + 1] = 1; |
2080 | rbuf[num + 3] = sat_model_serial_desc_len; |
2081 | num += 4; |
2082 | memcpy(rbuf + num, "ATA ", 8); |
2083 | num += 8; |
2084 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_PROD, |
2085 | ATA_ID_PROD_LEN); |
2086 | num += ATA_ID_PROD_LEN; |
2087 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_SERNO, |
2088 | ATA_ID_SERNO_LEN); |
2089 | num += ATA_ID_SERNO_LEN; |
2090 | |
2091 | if (ata_id_has_wwn(args->id)) { |
2092 | /* SAT defined lu world wide name */ |
2093 | /* piv=0, assoc=lu, code_set=binary, designator=NAA */ |
2094 | rbuf[num + 0] = 1; |
2095 | rbuf[num + 1] = 3; |
2096 | rbuf[num + 3] = ATA_ID_WWN_LEN; |
2097 | num += 4; |
2098 | ata_id_string(args->id, (unsigned char *) rbuf + num, |
2099 | ATA_ID_WWN, ATA_ID_WWN_LEN); |
2100 | num += ATA_ID_WWN_LEN; |
2101 | } |
2102 | rbuf[3] = num - 4; /* page len (assume less than 256 bytes) */ |
2103 | return 0; |
2104 | } |
2105 | |
2106 | /** |
2107 | * ata_scsiop_inq_89 - Simulate INQUIRY VPD page 89, ATA info |
2108 | * @args: device IDENTIFY data / SCSI command of interest. |
2109 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2110 | * |
2111 | * Yields SAT-specified ATA VPD page. |
2112 | * |
2113 | * LOCKING: |
2114 | * spin_lock_irqsave(host lock) |
2115 | */ |
2116 | static unsigned int ata_scsiop_inq_89(struct ata_scsi_args *args, u8 *rbuf) |
2117 | { |
2118 | struct ata_taskfile tf; |
2119 | |
2120 | memset(&tf, 0, sizeof(tf)); |
2121 | |
2122 | rbuf[1] = 0x89; /* our page code */ |
2123 | rbuf[2] = (0x238 >> 8); /* page size fixed at 238h */ |
2124 | rbuf[3] = (0x238 & 0xff); |
2125 | |
2126 | memcpy(&rbuf[8], "linux ", 8); |
2127 | memcpy(&rbuf[16], "libata ", 16); |
2128 | memcpy(&rbuf[32], DRV_VERSION, 4); |
2129 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV, 4); |
2130 | |
2131 | /* we don't store the ATA device signature, so we fake it */ |
2132 | |
2133 | tf.command = ATA_DRDY; /* really, this is Status reg */ |
2134 | tf.lbal = 0x1; |
2135 | tf.nsect = 0x1; |
2136 | |
2137 | ata_tf_to_fis(&tf, 0, 1, &rbuf[36]); /* TODO: PMP? */ |
2138 | rbuf[36] = 0x34; /* force D2H Reg FIS (34h) */ |
2139 | |
2140 | rbuf[56] = ATA_CMD_ID_ATA; |
2141 | |
2142 | memcpy(&rbuf[60], &args->id[0], 512); |
2143 | return 0; |
2144 | } |
2145 | |
2146 | static unsigned int ata_scsiop_inq_b0(struct ata_scsi_args *args, u8 *rbuf) |
2147 | { |
2148 | u16 min_io_sectors; |
2149 | |
2150 | rbuf[1] = 0xb0; |
2151 | rbuf[3] = 0x3c; /* required VPD size with unmap support */ |
2152 | |
2153 | /* |
2154 | * Optimal transfer length granularity. |
2155 | * |
2156 | * This is always one physical block, but for disks with a smaller |
2157 | * logical than physical sector size we need to figure out what the |
2158 | * latter is. |
2159 | */ |
2160 | min_io_sectors = 1 << ata_id_log2_per_physical_sector(args->id); |
2161 | put_unaligned_be16(min_io_sectors, &rbuf[6]); |
2162 | |
2163 | /* |
2164 | * Optimal unmap granularity. |
2165 | * |
2166 | * The ATA spec doesn't even know about a granularity or alignment |
2167 | * for the TRIM command. We can leave away most of the unmap related |
2168 | * VPD page entries, but we have specifify a granularity to signal |
2169 | * that we support some form of unmap - in thise case via WRITE SAME |
2170 | * with the unmap bit set. |
2171 | */ |
2172 | if (ata_id_has_trim(args->id)) { |
2173 | put_unaligned_be64(65535 * 512 / 8, &rbuf[36]); |
2174 | put_unaligned_be32(1, &rbuf[28]); |
2175 | } |
2176 | |
2177 | return 0; |
2178 | } |
2179 | |
2180 | static unsigned int ata_scsiop_inq_b1(struct ata_scsi_args *args, u8 *rbuf) |
2181 | { |
2182 | int form_factor = ata_id_form_factor(args->id); |
2183 | int media_rotation_rate = ata_id_rotation_rate(args->id); |
2184 | |
2185 | rbuf[1] = 0xb1; |
2186 | rbuf[3] = 0x3c; |
2187 | rbuf[4] = media_rotation_rate >> 8; |
2188 | rbuf[5] = media_rotation_rate; |
2189 | rbuf[7] = form_factor; |
2190 | |
2191 | return 0; |
2192 | } |
2193 | |
2194 | static unsigned int ata_scsiop_inq_b2(struct ata_scsi_args *args, u8 *rbuf) |
2195 | { |
2196 | /* SCSI Thin Provisioning VPD page: SBC-3 rev 22 or later */ |
2197 | rbuf[1] = 0xb2; |
2198 | rbuf[3] = 0x4; |
2199 | rbuf[5] = 1 << 6; /* TPWS */ |
2200 | |
2201 | return 0; |
2202 | } |
2203 | |
2204 | /** |
2205 | * ata_scsiop_noop - Command handler that simply returns success. |
2206 | * @args: device IDENTIFY data / SCSI command of interest. |
2207 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2208 | * |
2209 | * No operation. Simply returns success to caller, to indicate |
2210 | * that the caller should successfully complete this SCSI command. |
2211 | * |
2212 | * LOCKING: |
2213 | * spin_lock_irqsave(host lock) |
2214 | */ |
2215 | static unsigned int ata_scsiop_noop(struct ata_scsi_args *args, u8 *rbuf) |
2216 | { |
2217 | VPRINTK("ENTER\n"); |
2218 | return 0; |
2219 | } |
2220 | |
2221 | /** |
2222 | * modecpy - Prepare response for MODE SENSE |
2223 | * @dest: output buffer |
2224 | * @src: data being copied |
2225 | * @n: length of mode page |
2226 | * @changeable: whether changeable parameters are requested |
2227 | * |
2228 | * Generate a generic MODE SENSE page for either current or changeable |
2229 | * parameters. |
2230 | * |
2231 | * LOCKING: |
2232 | * None. |
2233 | */ |
2234 | static void modecpy(u8 *dest, const u8 *src, int n, bool changeable) |
2235 | { |
2236 | if (changeable) { |
2237 | memcpy(dest, src, 2); |
2238 | memset(dest + 2, 0, n - 2); |
2239 | } else { |
2240 | memcpy(dest, src, n); |
2241 | } |
2242 | } |
2243 | |
2244 | /** |
2245 | * ata_msense_caching - Simulate MODE SENSE caching info page |
2246 | * @id: device IDENTIFY data |
2247 | * @buf: output buffer |
2248 | * @changeable: whether changeable parameters are requested |
2249 | * |
2250 | * Generate a caching info page, which conditionally indicates |
2251 | * write caching to the SCSI layer, depending on device |
2252 | * capabilities. |
2253 | * |
2254 | * LOCKING: |
2255 | * None. |
2256 | */ |
2257 | static unsigned int ata_msense_caching(u16 *id, u8 *buf, bool changeable) |
2258 | { |
2259 | modecpy(buf, def_cache_mpage, sizeof(def_cache_mpage), changeable); |
2260 | if (changeable || ata_id_wcache_enabled(id)) |
2261 | buf[2] |= (1 << 2); /* write cache enable */ |
2262 | if (!changeable && !ata_id_rahead_enabled(id)) |
2263 | buf[12] |= (1 << 5); /* disable read ahead */ |
2264 | return sizeof(def_cache_mpage); |
2265 | } |
2266 | |
2267 | /** |
2268 | * ata_msense_ctl_mode - Simulate MODE SENSE control mode page |
2269 | * @buf: output buffer |
2270 | * @changeable: whether changeable parameters are requested |
2271 | * |
2272 | * Generate a generic MODE SENSE control mode page. |
2273 | * |
2274 | * LOCKING: |
2275 | * None. |
2276 | */ |
2277 | static unsigned int ata_msense_ctl_mode(u8 *buf, bool changeable) |
2278 | { |
2279 | modecpy(buf, def_control_mpage, sizeof(def_control_mpage), changeable); |
2280 | return sizeof(def_control_mpage); |
2281 | } |
2282 | |
2283 | /** |
2284 | * ata_msense_rw_recovery - Simulate MODE SENSE r/w error recovery page |
2285 | * @buf: output buffer |
2286 | * @changeable: whether changeable parameters are requested |
2287 | * |
2288 | * Generate a generic MODE SENSE r/w error recovery page. |
2289 | * |
2290 | * LOCKING: |
2291 | * None. |
2292 | */ |
2293 | static unsigned int ata_msense_rw_recovery(u8 *buf, bool changeable) |
2294 | { |
2295 | modecpy(buf, def_rw_recovery_mpage, sizeof(def_rw_recovery_mpage), |
2296 | changeable); |
2297 | return sizeof(def_rw_recovery_mpage); |
2298 | } |
2299 | |
2300 | /* |
2301 | * We can turn this into a real blacklist if it's needed, for now just |
2302 | * blacklist any Maxtor BANC1G10 revision firmware |
2303 | */ |
2304 | static int ata_dev_supports_fua(u16 *id) |
2305 | { |
2306 | unsigned char model[ATA_ID_PROD_LEN + 1], fw[ATA_ID_FW_REV_LEN + 1]; |
2307 | |
2308 | if (!libata_fua) |
2309 | return 0; |
2310 | if (!ata_id_has_fua(id)) |
2311 | return 0; |
2312 | |
2313 | ata_id_c_string(id, model, ATA_ID_PROD, sizeof(model)); |
2314 | ata_id_c_string(id, fw, ATA_ID_FW_REV, sizeof(fw)); |
2315 | |
2316 | if (strcmp(model, "Maxtor")) |
2317 | return 1; |
2318 | if (strcmp(fw, "BANC1G10")) |
2319 | return 1; |
2320 | |
2321 | return 0; /* blacklisted */ |
2322 | } |
2323 | |
2324 | /** |
2325 | * ata_scsiop_mode_sense - Simulate MODE SENSE 6, 10 commands |
2326 | * @args: device IDENTIFY data / SCSI command of interest. |
2327 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2328 | * |
2329 | * Simulate MODE SENSE commands. Assume this is invoked for direct |
2330 | * access devices (e.g. disks) only. There should be no block |
2331 | * descriptor for other device types. |
2332 | * |
2333 | * LOCKING: |
2334 | * spin_lock_irqsave(host lock) |
2335 | */ |
2336 | static unsigned int ata_scsiop_mode_sense(struct ata_scsi_args *args, u8 *rbuf) |
2337 | { |
2338 | struct ata_device *dev = args->dev; |
2339 | u8 *scsicmd = args->cmd->cmnd, *p = rbuf; |
2340 | const u8 sat_blk_desc[] = { |
2341 | 0, 0, 0, 0, /* number of blocks: sat unspecified */ |
2342 | 0, |
2343 | 0, 0x2, 0x0 /* block length: 512 bytes */ |
2344 | }; |
2345 | u8 pg, spg; |
2346 | unsigned int ebd, page_control, six_byte; |
2347 | u8 dpofua; |
2348 | |
2349 | VPRINTK("ENTER\n"); |
2350 | |
2351 | six_byte = (scsicmd[0] == MODE_SENSE); |
2352 | ebd = !(scsicmd[1] & 0x8); /* dbd bit inverted == edb */ |
2353 | /* |
2354 | * LLBA bit in msense(10) ignored (compliant) |
2355 | */ |
2356 | |
2357 | page_control = scsicmd[2] >> 6; |
2358 | switch (page_control) { |
2359 | case 0: /* current */ |
2360 | case 1: /* changeable */ |
2361 | case 2: /* defaults */ |
2362 | break; /* supported */ |
2363 | case 3: /* saved */ |
2364 | goto saving_not_supp; |
2365 | default: |
2366 | goto invalid_fld; |
2367 | } |
2368 | |
2369 | if (six_byte) |
2370 | p += 4 + (ebd ? 8 : 0); |
2371 | else |
2372 | p += 8 + (ebd ? 8 : 0); |
2373 | |
2374 | pg = scsicmd[2] & 0x3f; |
2375 | spg = scsicmd[3]; |
2376 | /* |
2377 | * No mode subpages supported (yet) but asking for _all_ |
2378 | * subpages may be valid |
2379 | */ |
2380 | if (spg && (spg != ALL_SUB_MPAGES)) |
2381 | goto invalid_fld; |
2382 | |
2383 | switch(pg) { |
2384 | case RW_RECOVERY_MPAGE: |
2385 | p += ata_msense_rw_recovery(p, page_control == 1); |
2386 | break; |
2387 | |
2388 | case CACHE_MPAGE: |
2389 | p += ata_msense_caching(args->id, p, page_control == 1); |
2390 | break; |
2391 | |
2392 | case CONTROL_MPAGE: |
2393 | p += ata_msense_ctl_mode(p, page_control == 1); |
2394 | break; |
2395 | |
2396 | case ALL_MPAGES: |
2397 | p += ata_msense_rw_recovery(p, page_control == 1); |
2398 | p += ata_msense_caching(args->id, p, page_control == 1); |
2399 | p += ata_msense_ctl_mode(p, page_control == 1); |
2400 | break; |
2401 | |
2402 | default: /* invalid page code */ |
2403 | goto invalid_fld; |
2404 | } |
2405 | |
2406 | dpofua = 0; |
2407 | if (ata_dev_supports_fua(args->id) && (dev->flags & ATA_DFLAG_LBA48) && |
2408 | (!(dev->flags & ATA_DFLAG_PIO) || dev->multi_count)) |
2409 | dpofua = 1 << 4; |
2410 | |
2411 | if (six_byte) { |
2412 | rbuf[0] = p - rbuf - 1; |
2413 | rbuf[2] |= dpofua; |
2414 | if (ebd) { |
2415 | rbuf[3] = sizeof(sat_blk_desc); |
2416 | memcpy(rbuf + 4, sat_blk_desc, sizeof(sat_blk_desc)); |
2417 | } |
2418 | } else { |
2419 | unsigned int output_len = p - rbuf - 2; |
2420 | |
2421 | rbuf[0] = output_len >> 8; |
2422 | rbuf[1] = output_len; |
2423 | rbuf[3] |= dpofua; |
2424 | if (ebd) { |
2425 | rbuf[7] = sizeof(sat_blk_desc); |
2426 | memcpy(rbuf + 8, sat_blk_desc, sizeof(sat_blk_desc)); |
2427 | } |
2428 | } |
2429 | return 0; |
2430 | |
2431 | invalid_fld: |
2432 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x24, 0x0); |
2433 | /* "Invalid field in cbd" */ |
2434 | return 1; |
2435 | |
2436 | saving_not_supp: |
2437 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x39, 0x0); |
2438 | /* "Saving parameters not supported" */ |
2439 | return 1; |
2440 | } |
2441 | |
2442 | /** |
2443 | * ata_scsiop_read_cap - Simulate READ CAPACITY[ 16] commands |
2444 | * @args: device IDENTIFY data / SCSI command of interest. |
2445 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2446 | * |
2447 | * Simulate READ CAPACITY commands. |
2448 | * |
2449 | * LOCKING: |
2450 | * None. |
2451 | */ |
2452 | static unsigned int ata_scsiop_read_cap(struct ata_scsi_args *args, u8 *rbuf) |
2453 | { |
2454 | struct ata_device *dev = args->dev; |
2455 | u64 last_lba = dev->n_sectors - 1; /* LBA of the last block */ |
2456 | u32 sector_size; /* physical sector size in bytes */ |
2457 | u8 log2_per_phys; |
2458 | u16 lowest_aligned; |
2459 | |
2460 | sector_size = ata_id_logical_sector_size(dev->id); |
2461 | log2_per_phys = ata_id_log2_per_physical_sector(dev->id); |
2462 | lowest_aligned = ata_id_logical_sector_offset(dev->id, log2_per_phys); |
2463 | |
2464 | VPRINTK("ENTER\n"); |
2465 | |
2466 | if (args->cmd->cmnd[0] == READ_CAPACITY) { |
2467 | if (last_lba >= 0xffffffffULL) |
2468 | last_lba = 0xffffffff; |
2469 | |
2470 | /* sector count, 32-bit */ |
2471 | rbuf[0] = last_lba >> (8 * 3); |
2472 | rbuf[1] = last_lba >> (8 * 2); |
2473 | rbuf[2] = last_lba >> (8 * 1); |
2474 | rbuf[3] = last_lba; |
2475 | |
2476 | /* sector size */ |
2477 | rbuf[4] = sector_size >> (8 * 3); |
2478 | rbuf[5] = sector_size >> (8 * 2); |
2479 | rbuf[6] = sector_size >> (8 * 1); |
2480 | rbuf[7] = sector_size; |
2481 | } else { |
2482 | /* sector count, 64-bit */ |
2483 | rbuf[0] = last_lba >> (8 * 7); |
2484 | rbuf[1] = last_lba >> (8 * 6); |
2485 | rbuf[2] = last_lba >> (8 * 5); |
2486 | rbuf[3] = last_lba >> (8 * 4); |
2487 | rbuf[4] = last_lba >> (8 * 3); |
2488 | rbuf[5] = last_lba >> (8 * 2); |
2489 | rbuf[6] = last_lba >> (8 * 1); |
2490 | rbuf[7] = last_lba; |
2491 | |
2492 | /* sector size */ |
2493 | rbuf[ 8] = sector_size >> (8 * 3); |
2494 | rbuf[ 9] = sector_size >> (8 * 2); |
2495 | rbuf[10] = sector_size >> (8 * 1); |
2496 | rbuf[11] = sector_size; |
2497 | |
2498 | rbuf[12] = 0; |
2499 | rbuf[13] = log2_per_phys; |
2500 | rbuf[14] = (lowest_aligned >> 8) & 0x3f; |
2501 | rbuf[15] = lowest_aligned; |
2502 | |
2503 | if (ata_id_has_trim(args->id)) { |
2504 | rbuf[14] |= 0x80; /* TPE */ |
2505 | |
2506 | if (ata_id_has_zero_after_trim(args->id)) |
2507 | rbuf[14] |= 0x40; /* TPRZ */ |
2508 | } |
2509 | } |
2510 | |
2511 | return 0; |
2512 | } |
2513 | |
2514 | /** |
2515 | * ata_scsiop_report_luns - Simulate REPORT LUNS command |
2516 | * @args: device IDENTIFY data / SCSI command of interest. |
2517 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2518 | * |
2519 | * Simulate REPORT LUNS command. |
2520 | * |
2521 | * LOCKING: |
2522 | * spin_lock_irqsave(host lock) |
2523 | */ |
2524 | static unsigned int ata_scsiop_report_luns(struct ata_scsi_args *args, u8 *rbuf) |
2525 | { |
2526 | VPRINTK("ENTER\n"); |
2527 | rbuf[3] = 8; /* just one lun, LUN 0, size 8 bytes */ |
2528 | |
2529 | return 0; |
2530 | } |
2531 | |
2532 | static void atapi_sense_complete(struct ata_queued_cmd *qc) |
2533 | { |
2534 | if (qc->err_mask && ((qc->err_mask & AC_ERR_DEV) == 0)) { |
2535 | /* FIXME: not quite right; we don't want the |
2536 | * translation of taskfile registers into |
2537 | * a sense descriptors, since that's only |
2538 | * correct for ATA, not ATAPI |
2539 | */ |
2540 | ata_gen_passthru_sense(qc); |
2541 | } |
2542 | |
2543 | qc->scsidone(qc->scsicmd); |
2544 | ata_qc_free(qc); |
2545 | } |
2546 | |
2547 | /* is it pointless to prefer PIO for "safety reasons"? */ |
2548 | static inline int ata_pio_use_silly(struct ata_port *ap) |
2549 | { |
2550 | return (ap->flags & ATA_FLAG_PIO_DMA); |
2551 | } |
2552 | |
2553 | static void atapi_request_sense(struct ata_queued_cmd *qc) |
2554 | { |
2555 | struct ata_port *ap = qc->ap; |
2556 | struct scsi_cmnd *cmd = qc->scsicmd; |
2557 | |
2558 | DPRINTK("ATAPI request sense\n"); |
2559 | |
2560 | /* FIXME: is this needed? */ |
2561 | memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); |
2562 | |
2563 | #ifdef CONFIG_ATA_SFF |
2564 | if (ap->ops->sff_tf_read) |
2565 | ap->ops->sff_tf_read(ap, &qc->tf); |
2566 | #endif |
2567 | |
2568 | /* fill these in, for the case where they are -not- overwritten */ |
2569 | cmd->sense_buffer[0] = 0x70; |
2570 | cmd->sense_buffer[2] = qc->tf.feature >> 4; |
2571 | |
2572 | ata_qc_reinit(qc); |
2573 | |
2574 | /* setup sg table and init transfer direction */ |
2575 | sg_init_one(&qc->sgent, cmd->sense_buffer, SCSI_SENSE_BUFFERSIZE); |
2576 | ata_sg_init(qc, &qc->sgent, 1); |
2577 | qc->dma_dir = DMA_FROM_DEVICE; |
2578 | |
2579 | memset(&qc->cdb, 0, qc->dev->cdb_len); |
2580 | qc->cdb[0] = REQUEST_SENSE; |
2581 | qc->cdb[4] = SCSI_SENSE_BUFFERSIZE; |
2582 | |
2583 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
2584 | qc->tf.command = ATA_CMD_PACKET; |
2585 | |
2586 | if (ata_pio_use_silly(ap)) { |
2587 | qc->tf.protocol = ATAPI_PROT_DMA; |
2588 | qc->tf.feature |= ATAPI_PKT_DMA; |
2589 | } else { |
2590 | qc->tf.protocol = ATAPI_PROT_PIO; |
2591 | qc->tf.lbam = SCSI_SENSE_BUFFERSIZE; |
2592 | qc->tf.lbah = 0; |
2593 | } |
2594 | qc->nbytes = SCSI_SENSE_BUFFERSIZE; |
2595 | |
2596 | qc->complete_fn = atapi_sense_complete; |
2597 | |
2598 | ata_qc_issue(qc); |
2599 | |
2600 | DPRINTK("EXIT\n"); |
2601 | } |
2602 | |
2603 | static void atapi_qc_complete(struct ata_queued_cmd *qc) |
2604 | { |
2605 | struct scsi_cmnd *cmd = qc->scsicmd; |
2606 | unsigned int err_mask = qc->err_mask; |
2607 | |
2608 | VPRINTK("ENTER, err_mask 0x%X\n", err_mask); |
2609 | |
2610 | /* handle completion from new EH */ |
2611 | if (unlikely(qc->ap->ops->error_handler && |
2612 | (err_mask || qc->flags & ATA_QCFLAG_SENSE_VALID))) { |
2613 | |
2614 | if (!(qc->flags & ATA_QCFLAG_SENSE_VALID)) { |
2615 | /* FIXME: not quite right; we don't want the |
2616 | * translation of taskfile registers into a |
2617 | * sense descriptors, since that's only |
2618 | * correct for ATA, not ATAPI |
2619 | */ |
2620 | ata_gen_passthru_sense(qc); |
2621 | } |
2622 | |
2623 | /* SCSI EH automatically locks door if sdev->locked is |
2624 | * set. Sometimes door lock request continues to |
2625 | * fail, for example, when no media is present. This |
2626 | * creates a loop - SCSI EH issues door lock which |
2627 | * fails and gets invoked again to acquire sense data |
2628 | * for the failed command. |
2629 | * |
2630 | * If door lock fails, always clear sdev->locked to |
2631 | * avoid this infinite loop. |
2632 | * |
2633 | * This may happen before SCSI scan is complete. Make |
2634 | * sure qc->dev->sdev isn't NULL before dereferencing. |
2635 | */ |
2636 | if (qc->cdb[0] == ALLOW_MEDIUM_REMOVAL && qc->dev->sdev) |
2637 | qc->dev->sdev->locked = 0; |
2638 | |
2639 | qc->scsicmd->result = SAM_STAT_CHECK_CONDITION; |
2640 | qc->scsidone(cmd); |
2641 | ata_qc_free(qc); |
2642 | return; |
2643 | } |
2644 | |
2645 | /* successful completion or old EH failure path */ |
2646 | if (unlikely(err_mask & AC_ERR_DEV)) { |
2647 | cmd->result = SAM_STAT_CHECK_CONDITION; |
2648 | atapi_request_sense(qc); |
2649 | return; |
2650 | } else if (unlikely(err_mask)) { |
2651 | /* FIXME: not quite right; we don't want the |
2652 | * translation of taskfile registers into |
2653 | * a sense descriptors, since that's only |
2654 | * correct for ATA, not ATAPI |
2655 | */ |
2656 | ata_gen_passthru_sense(qc); |
2657 | } else { |
2658 | u8 *scsicmd = cmd->cmnd; |
2659 | |
2660 | if ((scsicmd[0] == INQUIRY) && ((scsicmd[1] & 0x03) == 0)) { |
2661 | unsigned long flags; |
2662 | u8 *buf; |
2663 | |
2664 | buf = ata_scsi_rbuf_get(cmd, true, &flags); |
2665 | |
2666 | /* ATAPI devices typically report zero for their SCSI version, |
2667 | * and sometimes deviate from the spec WRT response data |
2668 | * format. If SCSI version is reported as zero like normal, |
2669 | * then we make the following fixups: 1) Fake MMC-5 version, |
2670 | * to indicate to the Linux scsi midlayer this is a modern |
2671 | * device. 2) Ensure response data format / ATAPI information |
2672 | * are always correct. |
2673 | */ |
2674 | if (buf[2] == 0) { |
2675 | buf[2] = 0x5; |
2676 | buf[3] = 0x32; |
2677 | } |
2678 | |
2679 | ata_scsi_rbuf_put(cmd, true, &flags); |
2680 | } |
2681 | |
2682 | cmd->result = SAM_STAT_GOOD; |
2683 | } |
2684 | |
2685 | qc->scsidone(cmd); |
2686 | ata_qc_free(qc); |
2687 | } |
2688 | /** |
2689 | * atapi_xlat - Initialize PACKET taskfile |
2690 | * @qc: command structure to be initialized |
2691 | * |
2692 | * LOCKING: |
2693 | * spin_lock_irqsave(host lock) |
2694 | * |
2695 | * RETURNS: |
2696 | * Zero on success, non-zero on failure. |
2697 | */ |
2698 | static unsigned int atapi_xlat(struct ata_queued_cmd *qc) |
2699 | { |
2700 | struct scsi_cmnd *scmd = qc->scsicmd; |
2701 | struct ata_device *dev = qc->dev; |
2702 | int nodata = (scmd->sc_data_direction == DMA_NONE); |
2703 | int using_pio = !nodata && (dev->flags & ATA_DFLAG_PIO); |
2704 | unsigned int nbytes; |
2705 | |
2706 | memset(qc->cdb, 0, dev->cdb_len); |
2707 | memcpy(qc->cdb, scmd->cmnd, scmd->cmd_len); |
2708 | |
2709 | qc->complete_fn = atapi_qc_complete; |
2710 | |
2711 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
2712 | if (scmd->sc_data_direction == DMA_TO_DEVICE) { |
2713 | qc->tf.flags |= ATA_TFLAG_WRITE; |
2714 | DPRINTK("direction: write\n"); |
2715 | } |
2716 | |
2717 | qc->tf.command = ATA_CMD_PACKET; |
2718 | ata_qc_set_pc_nbytes(qc); |
2719 | |
2720 | /* check whether ATAPI DMA is safe */ |
2721 | if (!nodata && !using_pio && atapi_check_dma(qc)) |
2722 | using_pio = 1; |
2723 | |
2724 | /* Some controller variants snoop this value for Packet |
2725 | * transfers to do state machine and FIFO management. Thus we |
2726 | * want to set it properly, and for DMA where it is |
2727 | * effectively meaningless. |
2728 | */ |
2729 | nbytes = min(ata_qc_raw_nbytes(qc), (unsigned int)63 * 1024); |
2730 | |
2731 | /* Most ATAPI devices which honor transfer chunk size don't |
2732 | * behave according to the spec when odd chunk size which |
2733 | * matches the transfer length is specified. If the number of |
2734 | * bytes to transfer is 2n+1. According to the spec, what |
2735 | * should happen is to indicate that 2n+1 is going to be |
2736 | * transferred and transfer 2n+2 bytes where the last byte is |
2737 | * padding. |
2738 | * |
2739 | * In practice, this doesn't happen. ATAPI devices first |
2740 | * indicate and transfer 2n bytes and then indicate and |
2741 | * transfer 2 bytes where the last byte is padding. |
2742 | * |
2743 | * This inconsistency confuses several controllers which |
2744 | * perform PIO using DMA such as Intel AHCIs and sil3124/32. |
2745 | * These controllers use actual number of transferred bytes to |
2746 | * update DMA poitner and transfer of 4n+2 bytes make those |
2747 | * controller push DMA pointer by 4n+4 bytes because SATA data |
2748 | * FISes are aligned to 4 bytes. This causes data corruption |
2749 | * and buffer overrun. |
2750 | * |
2751 | * Always setting nbytes to even number solves this problem |
2752 | * because then ATAPI devices don't have to split data at 2n |
2753 | * boundaries. |
2754 | */ |
2755 | if (nbytes & 0x1) |
2756 | nbytes++; |
2757 | |
2758 | qc->tf.lbam = (nbytes & 0xFF); |
2759 | qc->tf.lbah = (nbytes >> 8); |
2760 | |
2761 | if (nodata) |
2762 | qc->tf.protocol = ATAPI_PROT_NODATA; |
2763 | else if (using_pio) |
2764 | qc->tf.protocol = ATAPI_PROT_PIO; |
2765 | else { |
2766 | /* DMA data xfer */ |
2767 | qc->tf.protocol = ATAPI_PROT_DMA; |
2768 | qc->tf.feature |= ATAPI_PKT_DMA; |
2769 | |
2770 | if ((dev->flags & ATA_DFLAG_DMADIR) && |
2771 | (scmd->sc_data_direction != DMA_TO_DEVICE)) |
2772 | /* some SATA bridges need us to indicate data xfer direction */ |
2773 | qc->tf.feature |= ATAPI_DMADIR; |
2774 | } |
2775 | |
2776 | |
2777 | /* FIXME: We need to translate 0x05 READ_BLOCK_LIMITS to a MODE_SENSE |
2778 | as ATAPI tape drives don't get this right otherwise */ |
2779 | return 0; |
2780 | } |
2781 | |
2782 | static struct ata_device *ata_find_dev(struct ata_port *ap, int devno) |
2783 | { |
2784 | if (!sata_pmp_attached(ap)) { |
2785 | if (likely(devno < ata_link_max_devices(&ap->link))) |
2786 | return &ap->link.device[devno]; |
2787 | } else { |
2788 | if (likely(devno < ap->nr_pmp_links)) |
2789 | return &ap->pmp_link[devno].device[0]; |
2790 | } |
2791 | |
2792 | return NULL; |
2793 | } |
2794 | |
2795 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, |
2796 | const struct scsi_device *scsidev) |
2797 | { |
2798 | int devno; |
2799 | |
2800 | /* skip commands not addressed to targets we simulate */ |
2801 | if (!sata_pmp_attached(ap)) { |
2802 | if (unlikely(scsidev->channel || scsidev->lun)) |
2803 | return NULL; |
2804 | devno = scsidev->id; |
2805 | } else { |
2806 | if (unlikely(scsidev->id || scsidev->lun)) |
2807 | return NULL; |
2808 | devno = scsidev->channel; |
2809 | } |
2810 | |
2811 | return ata_find_dev(ap, devno); |
2812 | } |
2813 | |
2814 | /** |
2815 | * ata_scsi_find_dev - lookup ata_device from scsi_cmnd |
2816 | * @ap: ATA port to which the device is attached |
2817 | * @scsidev: SCSI device from which we derive the ATA device |
2818 | * |
2819 | * Given various information provided in struct scsi_cmnd, |
2820 | * map that onto an ATA bus, and using that mapping |
2821 | * determine which ata_device is associated with the |
2822 | * SCSI command to be sent. |
2823 | * |
2824 | * LOCKING: |
2825 | * spin_lock_irqsave(host lock) |
2826 | * |
2827 | * RETURNS: |
2828 | * Associated ATA device, or %NULL if not found. |
2829 | */ |
2830 | static struct ata_device * |
2831 | ata_scsi_find_dev(struct ata_port *ap, const struct scsi_device *scsidev) |
2832 | { |
2833 | struct ata_device *dev = __ata_scsi_find_dev(ap, scsidev); |
2834 | |
2835 | if (unlikely(!dev || !ata_dev_enabled(dev))) |
2836 | return NULL; |
2837 | |
2838 | return dev; |
2839 | } |
2840 | |
2841 | /* |
2842 | * ata_scsi_map_proto - Map pass-thru protocol value to taskfile value. |
2843 | * @byte1: Byte 1 from pass-thru CDB. |
2844 | * |
2845 | * RETURNS: |
2846 | * ATA_PROT_UNKNOWN if mapping failed/unimplemented, protocol otherwise. |
2847 | */ |
2848 | static u8 |
2849 | ata_scsi_map_proto(u8 byte1) |
2850 | { |
2851 | switch((byte1 & 0x1e) >> 1) { |
2852 | case 3: /* Non-data */ |
2853 | return ATA_PROT_NODATA; |
2854 | |
2855 | case 6: /* DMA */ |
2856 | case 10: /* UDMA Data-in */ |
2857 | case 11: /* UDMA Data-Out */ |
2858 | return ATA_PROT_DMA; |
2859 | |
2860 | case 4: /* PIO Data-in */ |
2861 | case 5: /* PIO Data-out */ |
2862 | return ATA_PROT_PIO; |
2863 | |
2864 | case 0: /* Hard Reset */ |
2865 | case 1: /* SRST */ |
2866 | case 8: /* Device Diagnostic */ |
2867 | case 9: /* Device Reset */ |
2868 | case 7: /* DMA Queued */ |
2869 | case 12: /* FPDMA */ |
2870 | case 15: /* Return Response Info */ |
2871 | default: /* Reserved */ |
2872 | break; |
2873 | } |
2874 | |
2875 | return ATA_PROT_UNKNOWN; |
2876 | } |
2877 | |
2878 | /** |
2879 | * ata_scsi_pass_thru - convert ATA pass-thru CDB to taskfile |
2880 | * @qc: command structure to be initialized |
2881 | * |
2882 | * Handles either 12 or 16-byte versions of the CDB. |
2883 | * |
2884 | * RETURNS: |
2885 | * Zero on success, non-zero on failure. |
2886 | */ |
2887 | static unsigned int ata_scsi_pass_thru(struct ata_queued_cmd *qc) |
2888 | { |
2889 | struct ata_taskfile *tf = &(qc->tf); |
2890 | struct scsi_cmnd *scmd = qc->scsicmd; |
2891 | struct ata_device *dev = qc->dev; |
2892 | const u8 *cdb = scmd->cmnd; |
2893 | |
2894 | if ((tf->protocol = ata_scsi_map_proto(cdb[1])) == ATA_PROT_UNKNOWN) |
2895 | goto invalid_fld; |
2896 | |
2897 | /* |
2898 | * 12 and 16 byte CDBs use different offsets to |
2899 | * provide the various register values. |
2900 | */ |
2901 | if (cdb[0] == ATA_16) { |
2902 | /* |
2903 | * 16-byte CDB - may contain extended commands. |
2904 | * |
2905 | * If that is the case, copy the upper byte register values. |
2906 | */ |
2907 | if (cdb[1] & 0x01) { |
2908 | tf->hob_feature = cdb[3]; |
2909 | tf->hob_nsect = cdb[5]; |
2910 | tf->hob_lbal = cdb[7]; |
2911 | tf->hob_lbam = cdb[9]; |
2912 | tf->hob_lbah = cdb[11]; |
2913 | tf->flags |= ATA_TFLAG_LBA48; |
2914 | } else |
2915 | tf->flags &= ~ATA_TFLAG_LBA48; |
2916 | |
2917 | /* |
2918 | * Always copy low byte, device and command registers. |
2919 | */ |
2920 | tf->feature = cdb[4]; |
2921 | tf->nsect = cdb[6]; |
2922 | tf->lbal = cdb[8]; |
2923 | tf->lbam = cdb[10]; |
2924 | tf->lbah = cdb[12]; |
2925 | tf->device = cdb[13]; |
2926 | tf->command = cdb[14]; |
2927 | } else { |
2928 | /* |
2929 | * 12-byte CDB - incapable of extended commands. |
2930 | */ |
2931 | tf->flags &= ~ATA_TFLAG_LBA48; |
2932 | |
2933 | tf->feature = cdb[3]; |
2934 | tf->nsect = cdb[4]; |
2935 | tf->lbal = cdb[5]; |
2936 | tf->lbam = cdb[6]; |
2937 | tf->lbah = cdb[7]; |
2938 | tf->device = cdb[8]; |
2939 | tf->command = cdb[9]; |
2940 | } |
2941 | |
2942 | /* enforce correct master/slave bit */ |
2943 | tf->device = dev->devno ? |
2944 | tf->device | ATA_DEV1 : tf->device & ~ATA_DEV1; |
2945 | |
2946 | switch (tf->command) { |
2947 | /* READ/WRITE LONG use a non-standard sect_size */ |
2948 | case ATA_CMD_READ_LONG: |
2949 | case ATA_CMD_READ_LONG_ONCE: |
2950 | case ATA_CMD_WRITE_LONG: |
2951 | case ATA_CMD_WRITE_LONG_ONCE: |
2952 | if (tf->protocol != ATA_PROT_PIO || tf->nsect != 1) |
2953 | goto invalid_fld; |
2954 | qc->sect_size = scsi_bufflen(scmd); |
2955 | break; |
2956 | |
2957 | /* commands using reported Logical Block size (e.g. 512 or 4K) */ |
2958 | case ATA_CMD_CFA_WRITE_NE: |
2959 | case ATA_CMD_CFA_TRANS_SECT: |
2960 | case ATA_CMD_CFA_WRITE_MULT_NE: |
2961 | /* XXX: case ATA_CMD_CFA_WRITE_SECTORS_WITHOUT_ERASE: */ |
2962 | case ATA_CMD_READ: |
2963 | case ATA_CMD_READ_EXT: |
2964 | case ATA_CMD_READ_QUEUED: |
2965 | /* XXX: case ATA_CMD_READ_QUEUED_EXT: */ |
2966 | case ATA_CMD_FPDMA_READ: |
2967 | case ATA_CMD_READ_MULTI: |
2968 | case ATA_CMD_READ_MULTI_EXT: |
2969 | case ATA_CMD_PIO_READ: |
2970 | case ATA_CMD_PIO_READ_EXT: |
2971 | case ATA_CMD_READ_STREAM_DMA_EXT: |
2972 | case ATA_CMD_READ_STREAM_EXT: |
2973 | case ATA_CMD_VERIFY: |
2974 | case ATA_CMD_VERIFY_EXT: |
2975 | case ATA_CMD_WRITE: |
2976 | case ATA_CMD_WRITE_EXT: |
2977 | case ATA_CMD_WRITE_FUA_EXT: |
2978 | case ATA_CMD_WRITE_QUEUED: |
2979 | case ATA_CMD_WRITE_QUEUED_FUA_EXT: |
2980 | case ATA_CMD_FPDMA_WRITE: |
2981 | case ATA_CMD_WRITE_MULTI: |
2982 | case ATA_CMD_WRITE_MULTI_EXT: |
2983 | case ATA_CMD_WRITE_MULTI_FUA_EXT: |
2984 | case ATA_CMD_PIO_WRITE: |
2985 | case ATA_CMD_PIO_WRITE_EXT: |
2986 | case ATA_CMD_WRITE_STREAM_DMA_EXT: |
2987 | case ATA_CMD_WRITE_STREAM_EXT: |
2988 | qc->sect_size = scmd->device->sector_size; |
2989 | break; |
2990 | |
2991 | /* Everything else uses 512 byte "sectors" */ |
2992 | default: |
2993 | qc->sect_size = ATA_SECT_SIZE; |
2994 | } |
2995 | |
2996 | /* |
2997 | * Set flags so that all registers will be written, pass on |
2998 | * write indication (used for PIO/DMA setup), result TF is |
2999 | * copied back and we don't whine too much about its failure. |
3000 | */ |
3001 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
3002 | if (scmd->sc_data_direction == DMA_TO_DEVICE) |
3003 | tf->flags |= ATA_TFLAG_WRITE; |
3004 | |
3005 | qc->flags |= ATA_QCFLAG_RESULT_TF | ATA_QCFLAG_QUIET; |
3006 | |
3007 | /* |
3008 | * Set transfer length. |
3009 | * |
3010 | * TODO: find out if we need to do more here to |
3011 | * cover scatter/gather case. |
3012 | */ |
3013 | ata_qc_set_pc_nbytes(qc); |
3014 | |
3015 | /* We may not issue DMA commands if no DMA mode is set */ |
3016 | if (tf->protocol == ATA_PROT_DMA && dev->dma_mode == 0) |
3017 | goto invalid_fld; |
3018 | |
3019 | /* sanity check for pio multi commands */ |
3020 | if ((cdb[1] & 0xe0) && !is_multi_taskfile(tf)) |
3021 | goto invalid_fld; |
3022 | |
3023 | if (is_multi_taskfile(tf)) { |
3024 | unsigned int multi_count = 1 << (cdb[1] >> 5); |
3025 | |
3026 | /* compare the passed through multi_count |
3027 | * with the cached multi_count of libata |
3028 | */ |
3029 | if (multi_count != dev->multi_count) |
3030 | ata_dev_warn(dev, "invalid multi_count %u ignored\n", |
3031 | multi_count); |
3032 | } |
3033 | |
3034 | /* |
3035 | * Filter SET_FEATURES - XFER MODE command -- otherwise, |
3036 | * SET_FEATURES - XFER MODE must be preceded/succeeded |
3037 | * by an update to hardware-specific registers for each |
3038 | * controller (i.e. the reason for ->set_piomode(), |
3039 | * ->set_dmamode(), and ->post_set_mode() hooks). |
3040 | */ |
3041 | if (tf->command == ATA_CMD_SET_FEATURES && |
3042 | tf->feature == SETFEATURES_XFER) |
3043 | goto invalid_fld; |
3044 | |
3045 | /* |
3046 | * Filter TPM commands by default. These provide an |
3047 | * essentially uncontrolled encrypted "back door" between |
3048 | * applications and the disk. Set libata.allow_tpm=1 if you |
3049 | * have a real reason for wanting to use them. This ensures |
3050 | * that installed software cannot easily mess stuff up without |
3051 | * user intent. DVR type users will probably ship with this enabled |
3052 | * for movie content management. |
3053 | * |
3054 | * Note that for ATA8 we can issue a DCS change and DCS freeze lock |
3055 | * for this and should do in future but that it is not sufficient as |
3056 | * DCS is an optional feature set. Thus we also do the software filter |
3057 | * so that we comply with the TC consortium stated goal that the user |
3058 | * can turn off TC features of their system. |
3059 | */ |
3060 | if (tf->command >= 0x5C && tf->command <= 0x5F && !libata_allow_tpm) |
3061 | goto invalid_fld; |
3062 | |
3063 | return 0; |
3064 | |
3065 | invalid_fld: |
3066 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); |
3067 | /* "Invalid field in cdb" */ |
3068 | return 1; |
3069 | } |
3070 | |
3071 | static unsigned int ata_scsi_write_same_xlat(struct ata_queued_cmd *qc) |
3072 | { |
3073 | struct ata_taskfile *tf = &qc->tf; |
3074 | struct scsi_cmnd *scmd = qc->scsicmd; |
3075 | struct ata_device *dev = qc->dev; |
3076 | const u8 *cdb = scmd->cmnd; |
3077 | u64 block; |
3078 | u32 n_block; |
3079 | u32 size; |
3080 | void *buf; |
3081 | |
3082 | /* we may not issue DMA commands if no DMA mode is set */ |
3083 | if (unlikely(!dev->dma_mode)) |
3084 | goto invalid_fld; |
3085 | |
3086 | if (unlikely(scmd->cmd_len < 16)) |
3087 | goto invalid_fld; |
3088 | scsi_16_lba_len(cdb, &block, &n_block); |
3089 | |
3090 | /* for now we only support WRITE SAME with the unmap bit set */ |
3091 | if (unlikely(!(cdb[1] & 0x8))) |
3092 | goto invalid_fld; |
3093 | |
3094 | /* |
3095 | * WRITE SAME always has a sector sized buffer as payload, this |
3096 | * should never be a multiple entry S/G list. |
3097 | */ |
3098 | if (!scsi_sg_count(scmd)) |
3099 | goto invalid_fld; |
3100 | |
3101 | buf = page_address(sg_page(scsi_sglist(scmd))); |
3102 | size = ata_set_lba_range_entries(buf, 512, block, n_block); |
3103 | |
3104 | tf->protocol = ATA_PROT_DMA; |
3105 | tf->hob_feature = 0; |
3106 | tf->feature = ATA_DSM_TRIM; |
3107 | tf->hob_nsect = (size / 512) >> 8; |
3108 | tf->nsect = size / 512; |
3109 | tf->command = ATA_CMD_DSM; |
3110 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE | ATA_TFLAG_LBA48 | |
3111 | ATA_TFLAG_WRITE; |
3112 | |
3113 | ata_qc_set_pc_nbytes(qc); |
3114 | |
3115 | return 0; |
3116 | |
3117 | invalid_fld: |
3118 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); |
3119 | /* "Invalid field in cdb" */ |
3120 | return 1; |
3121 | } |
3122 | |
3123 | /** |
3124 | * ata_mselect_caching - Simulate MODE SELECT for caching info page |
3125 | * @qc: Storage for translated ATA taskfile |
3126 | * @buf: input buffer |
3127 | * @len: number of valid bytes in the input buffer |
3128 | * |
3129 | * Prepare a taskfile to modify caching information for the device. |
3130 | * |
3131 | * LOCKING: |
3132 | * None. |
3133 | */ |
3134 | static int ata_mselect_caching(struct ata_queued_cmd *qc, |
3135 | const u8 *buf, int len) |
3136 | { |
3137 | struct ata_taskfile *tf = &qc->tf; |
3138 | struct ata_device *dev = qc->dev; |
3139 | char mpage[CACHE_MPAGE_LEN]; |
3140 | u8 wce; |
3141 | |
3142 | /* |
3143 | * The first two bytes of def_cache_mpage are a header, so offsets |
3144 | * in mpage are off by 2 compared to buf. Same for len. |
3145 | */ |
3146 | |
3147 | if (len != CACHE_MPAGE_LEN - 2) |
3148 | return -EINVAL; |
3149 | |
3150 | wce = buf[0] & (1 << 2); |
3151 | |
3152 | /* |
3153 | * Check that read-only bits are not modified. |
3154 | */ |
3155 | ata_msense_caching(dev->id, mpage, false); |
3156 | mpage[2] &= ~(1 << 2); |
3157 | mpage[2] |= wce; |
3158 | if (memcmp(mpage + 2, buf, CACHE_MPAGE_LEN - 2) != 0) |
3159 | return -EINVAL; |
3160 | |
3161 | tf->flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; |
3162 | tf->protocol = ATA_PROT_NODATA; |
3163 | tf->nsect = 0; |
3164 | tf->command = ATA_CMD_SET_FEATURES; |
3165 | tf->feature = wce ? SETFEATURES_WC_ON : SETFEATURES_WC_OFF; |
3166 | return 0; |
3167 | } |
3168 | |
3169 | /** |
3170 | * ata_scsiop_mode_select - Simulate MODE SELECT 6, 10 commands |
3171 | * @qc: Storage for translated ATA taskfile |
3172 | * |
3173 | * Converts a MODE SELECT command to an ATA SET FEATURES taskfile. |
3174 | * Assume this is invoked for direct access devices (e.g. disks) only. |
3175 | * There should be no block descriptor for other device types. |
3176 | * |
3177 | * LOCKING: |
3178 | * spin_lock_irqsave(host lock) |
3179 | */ |
3180 | static unsigned int ata_scsi_mode_select_xlat(struct ata_queued_cmd *qc) |
3181 | { |
3182 | struct scsi_cmnd *scmd = qc->scsicmd; |
3183 | const u8 *cdb = scmd->cmnd; |
3184 | const u8 *p; |
3185 | u8 pg, spg; |
3186 | unsigned six_byte, pg_len, hdr_len, bd_len; |
3187 | int len; |
3188 | |
3189 | VPRINTK("ENTER\n"); |
3190 | |
3191 | six_byte = (cdb[0] == MODE_SELECT); |
3192 | if (six_byte) { |
3193 | if (scmd->cmd_len < 5) |
3194 | goto invalid_fld; |
3195 | |
3196 | len = cdb[4]; |
3197 | hdr_len = 4; |
3198 | } else { |
3199 | if (scmd->cmd_len < 9) |
3200 | goto invalid_fld; |
3201 | |
3202 | len = (cdb[7] << 8) + cdb[8]; |
3203 | hdr_len = 8; |
3204 | } |
3205 | |
3206 | /* We only support PF=1, SP=0. */ |
3207 | if ((cdb[1] & 0x11) != 0x10) |
3208 | goto invalid_fld; |
3209 | |
3210 | /* Test early for possible overrun. */ |
3211 | if (!scsi_sg_count(scmd) || scsi_sglist(scmd)->length < len) |
3212 | goto invalid_param_len; |
3213 | |
3214 | p = page_address(sg_page(scsi_sglist(scmd))); |
3215 | |
3216 | /* Move past header and block descriptors. */ |
3217 | if (len < hdr_len) |
3218 | goto invalid_param_len; |
3219 | |
3220 | if (six_byte) |
3221 | bd_len = p[3]; |
3222 | else |
3223 | bd_len = (p[6] << 8) + p[7]; |
3224 | |
3225 | len -= hdr_len; |
3226 | p += hdr_len; |
3227 | if (len < bd_len) |
3228 | goto invalid_param_len; |
3229 | if (bd_len != 0 && bd_len != 8) |
3230 | goto invalid_param; |
3231 | |
3232 | len -= bd_len; |
3233 | p += bd_len; |
3234 | if (len == 0) |
3235 | goto skip; |
3236 | |
3237 | /* Parse both possible formats for the mode page headers. */ |
3238 | pg = p[0] & 0x3f; |
3239 | if (p[0] & 0x40) { |
3240 | if (len < 4) |
3241 | goto invalid_param_len; |
3242 | |
3243 | spg = p[1]; |
3244 | pg_len = (p[2] << 8) | p[3]; |
3245 | p += 4; |
3246 | len -= 4; |
3247 | } else { |
3248 | if (len < 2) |
3249 | goto invalid_param_len; |
3250 | |
3251 | spg = 0; |
3252 | pg_len = p[1]; |
3253 | p += 2; |
3254 | len -= 2; |
3255 | } |
3256 | |
3257 | /* |
3258 | * No mode subpages supported (yet) but asking for _all_ |
3259 | * subpages may be valid |
3260 | */ |
3261 | if (spg && (spg != ALL_SUB_MPAGES)) |
3262 | goto invalid_param; |
3263 | if (pg_len > len) |
3264 | goto invalid_param_len; |
3265 | |
3266 | switch (pg) { |
3267 | case CACHE_MPAGE: |
3268 | if (ata_mselect_caching(qc, p, pg_len) < 0) |
3269 | goto invalid_param; |
3270 | break; |
3271 | |
3272 | default: /* invalid page code */ |
3273 | goto invalid_param; |
3274 | } |
3275 | |
3276 | /* |
3277 | * Only one page has changeable data, so we only support setting one |
3278 | * page at a time. |
3279 | */ |
3280 | if (len > pg_len) |
3281 | goto invalid_param; |
3282 | |
3283 | return 0; |
3284 | |
3285 | invalid_fld: |
3286 | /* "Invalid field in CDB" */ |
3287 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
3288 | return 1; |
3289 | |
3290 | invalid_param: |
3291 | /* "Invalid field in parameter list" */ |
3292 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x26, 0x0); |
3293 | return 1; |
3294 | |
3295 | invalid_param_len: |
3296 | /* "Parameter list length error" */ |
3297 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x1a, 0x0); |
3298 | return 1; |
3299 | |
3300 | skip: |
3301 | scmd->result = SAM_STAT_GOOD; |
3302 | return 1; |
3303 | } |
3304 | |
3305 | /** |
3306 | * ata_get_xlat_func - check if SCSI to ATA translation is possible |
3307 | * @dev: ATA device |
3308 | * @cmd: SCSI command opcode to consider |
3309 | * |
3310 | * Look up the SCSI command given, and determine whether the |
3311 | * SCSI command is to be translated or simulated. |
3312 | * |
3313 | * RETURNS: |
3314 | * Pointer to translation function if possible, %NULL if not. |
3315 | */ |
3316 | |
3317 | static inline ata_xlat_func_t ata_get_xlat_func(struct ata_device *dev, u8 cmd) |
3318 | { |
3319 | switch (cmd) { |
3320 | case READ_6: |
3321 | case READ_10: |
3322 | case READ_16: |
3323 | |
3324 | case WRITE_6: |
3325 | case WRITE_10: |
3326 | case WRITE_16: |
3327 | return ata_scsi_rw_xlat; |
3328 | |
3329 | case WRITE_SAME_16: |
3330 | return ata_scsi_write_same_xlat; |
3331 | |
3332 | case SYNCHRONIZE_CACHE: |
3333 | if (ata_try_flush_cache(dev)) |
3334 | return ata_scsi_flush_xlat; |
3335 | break; |
3336 | |
3337 | case VERIFY: |
3338 | case VERIFY_16: |
3339 | return ata_scsi_verify_xlat; |
3340 | |
3341 | case ATA_12: |
3342 | case ATA_16: |
3343 | return ata_scsi_pass_thru; |
3344 | |
3345 | case MODE_SELECT: |
3346 | case MODE_SELECT_10: |
3347 | return ata_scsi_mode_select_xlat; |
3348 | break; |
3349 | |
3350 | case START_STOP: |
3351 | return ata_scsi_start_stop_xlat; |
3352 | } |
3353 | |
3354 | return NULL; |
3355 | } |
3356 | |
3357 | /** |
3358 | * ata_scsi_dump_cdb - dump SCSI command contents to dmesg |
3359 | * @ap: ATA port to which the command was being sent |
3360 | * @cmd: SCSI command to dump |
3361 | * |
3362 | * Prints the contents of a SCSI command via printk(). |
3363 | */ |
3364 | |
3365 | static inline void ata_scsi_dump_cdb(struct ata_port *ap, |
3366 | struct scsi_cmnd *cmd) |
3367 | { |
3368 | #ifdef ATA_DEBUG |
3369 | struct scsi_device *scsidev = cmd->device; |
3370 | u8 *scsicmd = cmd->cmnd; |
3371 | |
3372 | DPRINTK("CDB (%u:%d,%d,%d) %02x %02x %02x %02x %02x %02x %02x %02x %02x\n", |
3373 | ap->print_id, |
3374 | scsidev->channel, scsidev->id, scsidev->lun, |
3375 | scsicmd[0], scsicmd[1], scsicmd[2], scsicmd[3], |
3376 | scsicmd[4], scsicmd[5], scsicmd[6], scsicmd[7], |
3377 | scsicmd[8]); |
3378 | #endif |
3379 | } |
3380 | |
3381 | static inline int __ata_scsi_queuecmd(struct scsi_cmnd *scmd, |
3382 | struct ata_device *dev) |
3383 | { |
3384 | u8 scsi_op = scmd->cmnd[0]; |
3385 | ata_xlat_func_t xlat_func; |
3386 | int rc = 0; |
3387 | |
3388 | if (dev->class == ATA_DEV_ATA) { |
3389 | if (unlikely(!scmd->cmd_len || scmd->cmd_len > dev->cdb_len)) |
3390 | goto bad_cdb_len; |
3391 | |
3392 | xlat_func = ata_get_xlat_func(dev, scsi_op); |
3393 | } else { |
3394 | if (unlikely(!scmd->cmd_len)) |
3395 | goto bad_cdb_len; |
3396 | |
3397 | xlat_func = NULL; |
3398 | if (likely((scsi_op != ATA_16) || !atapi_passthru16)) { |
3399 | /* relay SCSI command to ATAPI device */ |
3400 | int len = COMMAND_SIZE(scsi_op); |
3401 | if (unlikely(len > scmd->cmd_len || len > dev->cdb_len)) |
3402 | goto bad_cdb_len; |
3403 | |
3404 | xlat_func = atapi_xlat; |
3405 | } else { |
3406 | /* ATA_16 passthru, treat as an ATA command */ |
3407 | if (unlikely(scmd->cmd_len > 16)) |
3408 | goto bad_cdb_len; |
3409 | |
3410 | xlat_func = ata_get_xlat_func(dev, scsi_op); |
3411 | } |
3412 | } |
3413 | |
3414 | if (xlat_func) |
3415 | rc = ata_scsi_translate(dev, scmd, xlat_func); |
3416 | else |
3417 | ata_scsi_simulate(dev, scmd); |
3418 | |
3419 | return rc; |
3420 | |
3421 | bad_cdb_len: |
3422 | DPRINTK("bad CDB len=%u, scsi_op=0x%02x, max=%u\n", |
3423 | scmd->cmd_len, scsi_op, dev->cdb_len); |
3424 | scmd->result = DID_ERROR << 16; |
3425 | scmd->scsi_done(scmd); |
3426 | return 0; |
3427 | } |
3428 | |
3429 | /** |
3430 | * ata_scsi_queuecmd - Issue SCSI cdb to libata-managed device |
3431 | * @shost: SCSI host of command to be sent |
3432 | * @cmd: SCSI command to be sent |
3433 | * |
3434 | * In some cases, this function translates SCSI commands into |
3435 | * ATA taskfiles, and queues the taskfiles to be sent to |
3436 | * hardware. In other cases, this function simulates a |
3437 | * SCSI device by evaluating and responding to certain |
3438 | * SCSI commands. This creates the overall effect of |
3439 | * ATA and ATAPI devices appearing as SCSI devices. |
3440 | * |
3441 | * LOCKING: |
3442 | * ATA host lock |
3443 | * |
3444 | * RETURNS: |
3445 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, |
3446 | * 0 otherwise. |
3447 | */ |
3448 | int ata_scsi_queuecmd(struct Scsi_Host *shost, struct scsi_cmnd *cmd) |
3449 | { |
3450 | struct ata_port *ap; |
3451 | struct ata_device *dev; |
3452 | struct scsi_device *scsidev = cmd->device; |
3453 | int rc = 0; |
3454 | unsigned long irq_flags; |
3455 | |
3456 | ap = ata_shost_to_port(shost); |
3457 | |
3458 | spin_lock_irqsave(ap->lock, irq_flags); |
3459 | |
3460 | ata_scsi_dump_cdb(ap, cmd); |
3461 | |
3462 | dev = ata_scsi_find_dev(ap, scsidev); |
3463 | if (likely(dev)) |
3464 | rc = __ata_scsi_queuecmd(cmd, dev); |
3465 | else { |
3466 | cmd->result = (DID_BAD_TARGET << 16); |
3467 | cmd->scsi_done(cmd); |
3468 | } |
3469 | |
3470 | spin_unlock_irqrestore(ap->lock, irq_flags); |
3471 | |
3472 | return rc; |
3473 | } |
3474 | |
3475 | /** |
3476 | * ata_scsi_simulate - simulate SCSI command on ATA device |
3477 | * @dev: the target device |
3478 | * @cmd: SCSI command being sent to device. |
3479 | * |
3480 | * Interprets and directly executes a select list of SCSI commands |
3481 | * that can be handled internally. |
3482 | * |
3483 | * LOCKING: |
3484 | * spin_lock_irqsave(host lock) |
3485 | */ |
3486 | |
3487 | void ata_scsi_simulate(struct ata_device *dev, struct scsi_cmnd *cmd) |
3488 | { |
3489 | struct ata_scsi_args args; |
3490 | const u8 *scsicmd = cmd->cmnd; |
3491 | u8 tmp8; |
3492 | |
3493 | args.dev = dev; |
3494 | args.id = dev->id; |
3495 | args.cmd = cmd; |
3496 | args.done = cmd->scsi_done; |
3497 | |
3498 | switch(scsicmd[0]) { |
3499 | /* TODO: worth improving? */ |
3500 | case FORMAT_UNIT: |
3501 | ata_scsi_invalid_field(cmd); |
3502 | break; |
3503 | |
3504 | case INQUIRY: |
3505 | if (scsicmd[1] & 2) /* is CmdDt set? */ |
3506 | ata_scsi_invalid_field(cmd); |
3507 | else if ((scsicmd[1] & 1) == 0) /* is EVPD clear? */ |
3508 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_std); |
3509 | else switch (scsicmd[2]) { |
3510 | case 0x00: |
3511 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_00); |
3512 | break; |
3513 | case 0x80: |
3514 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_80); |
3515 | break; |
3516 | case 0x83: |
3517 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_83); |
3518 | break; |
3519 | case 0x89: |
3520 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_89); |
3521 | break; |
3522 | case 0xb0: |
3523 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b0); |
3524 | break; |
3525 | case 0xb1: |
3526 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b1); |
3527 | break; |
3528 | case 0xb2: |
3529 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b2); |
3530 | break; |
3531 | default: |
3532 | ata_scsi_invalid_field(cmd); |
3533 | break; |
3534 | } |
3535 | break; |
3536 | |
3537 | case MODE_SENSE: |
3538 | case MODE_SENSE_10: |
3539 | ata_scsi_rbuf_fill(&args, ata_scsiop_mode_sense); |
3540 | break; |
3541 | |
3542 | case READ_CAPACITY: |
3543 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); |
3544 | break; |
3545 | |
3546 | case SERVICE_ACTION_IN: |
3547 | if ((scsicmd[1] & 0x1f) == SAI_READ_CAPACITY_16) |
3548 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); |
3549 | else |
3550 | ata_scsi_invalid_field(cmd); |
3551 | break; |
3552 | |
3553 | case REPORT_LUNS: |
3554 | ata_scsi_rbuf_fill(&args, ata_scsiop_report_luns); |
3555 | break; |
3556 | |
3557 | case REQUEST_SENSE: |
3558 | ata_scsi_set_sense(cmd, 0, 0, 0); |
3559 | cmd->result = (DRIVER_SENSE << 24); |
3560 | cmd->scsi_done(cmd); |
3561 | break; |
3562 | |
3563 | /* if we reach this, then writeback caching is disabled, |
3564 | * turning this into a no-op. |
3565 | */ |
3566 | case SYNCHRONIZE_CACHE: |
3567 | /* fall through */ |
3568 | |
3569 | /* no-op's, complete with success */ |
3570 | case REZERO_UNIT: |
3571 | case SEEK_6: |
3572 | case SEEK_10: |
3573 | case TEST_UNIT_READY: |
3574 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); |
3575 | break; |
3576 | |
3577 | case SEND_DIAGNOSTIC: |
3578 | tmp8 = scsicmd[1] & ~(1 << 3); |
3579 | if ((tmp8 == 0x4) && (!scsicmd[3]) && (!scsicmd[4])) |
3580 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); |
3581 | else |
3582 | ata_scsi_invalid_field(cmd); |
3583 | break; |
3584 | |
3585 | /* all other commands */ |
3586 | default: |
3587 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x20, 0x0); |
3588 | /* "Invalid command operation code" */ |
3589 | cmd->scsi_done(cmd); |
3590 | break; |
3591 | } |
3592 | } |
3593 | |
3594 | int ata_scsi_add_hosts(struct ata_host *host, struct scsi_host_template *sht) |
3595 | { |
3596 | int i, rc; |
3597 | |
3598 | for (i = 0; i < host->n_ports; i++) { |
3599 | struct ata_port *ap = host->ports[i]; |
3600 | struct Scsi_Host *shost; |
3601 | |
3602 | rc = -ENOMEM; |
3603 | shost = scsi_host_alloc(sht, sizeof(struct ata_port *)); |
3604 | if (!shost) |
3605 | goto err_alloc; |
3606 | |
3607 | shost->eh_noresume = 1; |
3608 | *(struct ata_port **)&shost->hostdata[0] = ap; |
3609 | ap->scsi_host = shost; |
3610 | |
3611 | shost->transportt = ata_scsi_transport_template; |
3612 | shost->unique_id = ap->print_id; |
3613 | shost->max_id = 16; |
3614 | shost->max_lun = 1; |
3615 | shost->max_channel = 1; |
3616 | shost->max_cmd_len = 16; |
3617 | |
3618 | /* Schedule policy is determined by ->qc_defer() |
3619 | * callback and it needs to see every deferred qc. |
3620 | * Set host_blocked to 1 to prevent SCSI midlayer from |
3621 | * automatically deferring requests. |
3622 | */ |
3623 | shost->max_host_blocked = 1; |
3624 | |
3625 | rc = scsi_add_host_with_dma(ap->scsi_host, |
3626 | &ap->tdev, ap->host->dev); |
3627 | if (rc) |
3628 | goto err_add; |
3629 | } |
3630 | |
3631 | return 0; |
3632 | |
3633 | err_add: |
3634 | scsi_host_put(host->ports[i]->scsi_host); |
3635 | err_alloc: |
3636 | while (--i >= 0) { |
3637 | struct Scsi_Host *shost = host->ports[i]->scsi_host; |
3638 | |
3639 | scsi_remove_host(shost); |
3640 | scsi_host_put(shost); |
3641 | } |
3642 | return rc; |
3643 | } |
3644 | |
3645 | void ata_scsi_scan_host(struct ata_port *ap, int sync) |
3646 | { |
3647 | int tries = 5; |
3648 | struct ata_device *last_failed_dev = NULL; |
3649 | struct ata_link *link; |
3650 | struct ata_device *dev; |
3651 | |
3652 | repeat: |
3653 | ata_for_each_link(link, ap, EDGE) { |
3654 | ata_for_each_dev(dev, link, ENABLED) { |
3655 | struct scsi_device *sdev; |
3656 | int channel = 0, id = 0; |
3657 | |
3658 | if (dev->sdev) |
3659 | continue; |
3660 | |
3661 | if (ata_is_host_link(link)) |
3662 | id = dev->devno; |
3663 | else |
3664 | channel = link->pmp; |
3665 | |
3666 | sdev = __scsi_add_device(ap->scsi_host, channel, id, 0, |
3667 | NULL); |
3668 | if (!IS_ERR(sdev)) { |
3669 | dev->sdev = sdev; |
3670 | scsi_device_put(sdev); |
3671 | ata_acpi_bind(dev); |
3672 | } else { |
3673 | dev->sdev = NULL; |
3674 | } |
3675 | } |
3676 | } |
3677 | |
3678 | /* If we scanned while EH was in progress or allocation |
3679 | * failure occurred, scan would have failed silently. Check |
3680 | * whether all devices are attached. |
3681 | */ |
3682 | ata_for_each_link(link, ap, EDGE) { |
3683 | ata_for_each_dev(dev, link, ENABLED) { |
3684 | if (!dev->sdev) |
3685 | goto exit_loop; |
3686 | } |
3687 | } |
3688 | exit_loop: |
3689 | if (!link) |
3690 | return; |
3691 | |
3692 | /* we're missing some SCSI devices */ |
3693 | if (sync) { |
3694 | /* If caller requested synchrnous scan && we've made |
3695 | * any progress, sleep briefly and repeat. |
3696 | */ |
3697 | if (dev != last_failed_dev) { |
3698 | msleep(100); |
3699 | last_failed_dev = dev; |
3700 | goto repeat; |
3701 | } |
3702 | |
3703 | /* We might be failing to detect boot device, give it |
3704 | * a few more chances. |
3705 | */ |
3706 | if (--tries) { |
3707 | msleep(100); |
3708 | goto repeat; |
3709 | } |
3710 | |
3711 | ata_port_err(ap, |
3712 | "WARNING: synchronous SCSI scan failed without making any progress, switching to async\n"); |
3713 | } |
3714 | |
3715 | queue_delayed_work(system_long_wq, &ap->hotplug_task, |
3716 | round_jiffies_relative(HZ)); |
3717 | } |
3718 | |
3719 | /** |
3720 | * ata_scsi_offline_dev - offline attached SCSI device |
3721 | * @dev: ATA device to offline attached SCSI device for |
3722 | * |
3723 | * This function is called from ata_eh_hotplug() and responsible |
3724 | * for taking the SCSI device attached to @dev offline. This |
3725 | * function is called with host lock which protects dev->sdev |
3726 | * against clearing. |
3727 | * |
3728 | * LOCKING: |
3729 | * spin_lock_irqsave(host lock) |
3730 | * |
3731 | * RETURNS: |
3732 | * 1 if attached SCSI device exists, 0 otherwise. |
3733 | */ |
3734 | int ata_scsi_offline_dev(struct ata_device *dev) |
3735 | { |
3736 | if (dev->sdev) { |
3737 | scsi_device_set_state(dev->sdev, SDEV_OFFLINE); |
3738 | return 1; |
3739 | } |
3740 | return 0; |
3741 | } |
3742 | |
3743 | /** |
3744 | * ata_scsi_remove_dev - remove attached SCSI device |
3745 | * @dev: ATA device to remove attached SCSI device for |
3746 | * |
3747 | * This function is called from ata_eh_scsi_hotplug() and |
3748 | * responsible for removing the SCSI device attached to @dev. |
3749 | * |
3750 | * LOCKING: |
3751 | * Kernel thread context (may sleep). |
3752 | */ |
3753 | static void ata_scsi_remove_dev(struct ata_device *dev) |
3754 | { |
3755 | struct ata_port *ap = dev->link->ap; |
3756 | struct scsi_device *sdev; |
3757 | unsigned long flags; |
3758 | |
3759 | /* Alas, we need to grab scan_mutex to ensure SCSI device |
3760 | * state doesn't change underneath us and thus |
3761 | * scsi_device_get() always succeeds. The mutex locking can |
3762 | * be removed if there is __scsi_device_get() interface which |
3763 | * increments reference counts regardless of device state. |
3764 | */ |
3765 | mutex_lock(&ap->scsi_host->scan_mutex); |
3766 | spin_lock_irqsave(ap->lock, flags); |
3767 | |
3768 | if (zpodd_dev_enabled(dev)) |
3769 | zpodd_exit(dev); |
3770 | ata_acpi_unbind(dev); |
3771 | |
3772 | /* clearing dev->sdev is protected by host lock */ |
3773 | sdev = dev->sdev; |
3774 | dev->sdev = NULL; |
3775 | |
3776 | if (sdev) { |
3777 | /* If user initiated unplug races with us, sdev can go |
3778 | * away underneath us after the host lock and |
3779 | * scan_mutex are released. Hold onto it. |
3780 | */ |
3781 | if (scsi_device_get(sdev) == 0) { |
3782 | /* The following ensures the attached sdev is |
3783 | * offline on return from ata_scsi_offline_dev() |
3784 | * regardless it wins or loses the race |
3785 | * against this function. |
3786 | */ |
3787 | scsi_device_set_state(sdev, SDEV_OFFLINE); |
3788 | } else { |
3789 | WARN_ON(1); |
3790 | sdev = NULL; |
3791 | } |
3792 | } |
3793 | |
3794 | spin_unlock_irqrestore(ap->lock, flags); |
3795 | mutex_unlock(&ap->scsi_host->scan_mutex); |
3796 | |
3797 | if (sdev) { |
3798 | ata_dev_info(dev, "detaching (SCSI %s)\n", |
3799 | dev_name(&sdev->sdev_gendev)); |
3800 | |
3801 | scsi_remove_device(sdev); |
3802 | scsi_device_put(sdev); |
3803 | } |
3804 | } |
3805 | |
3806 | static void ata_scsi_handle_link_detach(struct ata_link *link) |
3807 | { |
3808 | struct ata_port *ap = link->ap; |
3809 | struct ata_device *dev; |
3810 | |
3811 | ata_for_each_dev(dev, link, ALL) { |
3812 | unsigned long flags; |
3813 | |
3814 | if (!(dev->flags & ATA_DFLAG_DETACHED)) |
3815 | continue; |
3816 | |
3817 | spin_lock_irqsave(ap->lock, flags); |
3818 | dev->flags &= ~ATA_DFLAG_DETACHED; |
3819 | spin_unlock_irqrestore(ap->lock, flags); |
3820 | |
3821 | ata_scsi_remove_dev(dev); |
3822 | } |
3823 | } |
3824 | |
3825 | /** |
3826 | * ata_scsi_media_change_notify - send media change event |
3827 | * @dev: Pointer to the disk device with media change event |
3828 | * |
3829 | * Tell the block layer to send a media change notification |
3830 | * event. |
3831 | * |
3832 | * LOCKING: |
3833 | * spin_lock_irqsave(host lock) |
3834 | */ |
3835 | void ata_scsi_media_change_notify(struct ata_device *dev) |
3836 | { |
3837 | if (dev->sdev) |
3838 | sdev_evt_send_simple(dev->sdev, SDEV_EVT_MEDIA_CHANGE, |
3839 | GFP_ATOMIC); |
3840 | } |
3841 | |
3842 | /** |
3843 | * ata_scsi_hotplug - SCSI part of hotplug |
3844 | * @work: Pointer to ATA port to perform SCSI hotplug on |
3845 | * |
3846 | * Perform SCSI part of hotplug. It's executed from a separate |
3847 | * workqueue after EH completes. This is necessary because SCSI |
3848 | * hot plugging requires working EH and hot unplugging is |
3849 | * synchronized with hot plugging with a mutex. |
3850 | * |
3851 | * LOCKING: |
3852 | * Kernel thread context (may sleep). |
3853 | */ |
3854 | void ata_scsi_hotplug(struct work_struct *work) |
3855 | { |
3856 | struct ata_port *ap = |
3857 | container_of(work, struct ata_port, hotplug_task.work); |
3858 | int i; |
3859 | |
3860 | if (ap->pflags & ATA_PFLAG_UNLOADING) { |
3861 | DPRINTK("ENTER/EXIT - unloading\n"); |
3862 | return; |
3863 | } |
3864 | |
3865 | DPRINTK("ENTER\n"); |
3866 | mutex_lock(&ap->scsi_scan_mutex); |
3867 | |
3868 | /* Unplug detached devices. We cannot use link iterator here |
3869 | * because PMP links have to be scanned even if PMP is |
3870 | * currently not attached. Iterate manually. |
3871 | */ |
3872 | ata_scsi_handle_link_detach(&ap->link); |
3873 | if (ap->pmp_link) |
3874 | for (i = 0; i < SATA_PMP_MAX_PORTS; i++) |
3875 | ata_scsi_handle_link_detach(&ap->pmp_link[i]); |
3876 | |
3877 | /* scan for new ones */ |
3878 | ata_scsi_scan_host(ap, 0); |
3879 | |
3880 | mutex_unlock(&ap->scsi_scan_mutex); |
3881 | DPRINTK("EXIT\n"); |
3882 | } |
3883 | |
3884 | /** |
3885 | * ata_scsi_user_scan - indication for user-initiated bus scan |
3886 | * @shost: SCSI host to scan |
3887 | * @channel: Channel to scan |
3888 | * @id: ID to scan |
3889 | * @lun: LUN to scan |
3890 | * |
3891 | * This function is called when user explicitly requests bus |
3892 | * scan. Set probe pending flag and invoke EH. |
3893 | * |
3894 | * LOCKING: |
3895 | * SCSI layer (we don't care) |
3896 | * |
3897 | * RETURNS: |
3898 | * Zero. |
3899 | */ |
3900 | int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, |
3901 | unsigned int id, unsigned int lun) |
3902 | { |
3903 | struct ata_port *ap = ata_shost_to_port(shost); |
3904 | unsigned long flags; |
3905 | int devno, rc = 0; |
3906 | |
3907 | if (!ap->ops->error_handler) |
3908 | return -EOPNOTSUPP; |
3909 | |
3910 | if (lun != SCAN_WILD_CARD && lun) |
3911 | return -EINVAL; |
3912 | |
3913 | if (!sata_pmp_attached(ap)) { |
3914 | if (channel != SCAN_WILD_CARD && channel) |
3915 | return -EINVAL; |
3916 | devno = id; |
3917 | } else { |
3918 | if (id != SCAN_WILD_CARD && id) |
3919 | return -EINVAL; |
3920 | devno = channel; |
3921 | } |
3922 | |
3923 | spin_lock_irqsave(ap->lock, flags); |
3924 | |
3925 | if (devno == SCAN_WILD_CARD) { |
3926 | struct ata_link *link; |
3927 | |
3928 | ata_for_each_link(link, ap, EDGE) { |
3929 | struct ata_eh_info *ehi = &link->eh_info; |
3930 | ehi->probe_mask |= ATA_ALL_DEVICES; |
3931 | ehi->action |= ATA_EH_RESET; |
3932 | } |
3933 | } else { |
3934 | struct ata_device *dev = ata_find_dev(ap, devno); |
3935 | |
3936 | if (dev) { |
3937 | struct ata_eh_info *ehi = &dev->link->eh_info; |
3938 | ehi->probe_mask |= 1 << dev->devno; |
3939 | ehi->action |= ATA_EH_RESET; |
3940 | } else |
3941 | rc = -EINVAL; |
3942 | } |
3943 | |
3944 | if (rc == 0) { |
3945 | ata_port_schedule_eh(ap); |
3946 | spin_unlock_irqrestore(ap->lock, flags); |
3947 | ata_port_wait_eh(ap); |
3948 | } else |
3949 | spin_unlock_irqrestore(ap->lock, flags); |
3950 | |
3951 | return rc; |
3952 | } |
3953 | |
3954 | /** |
3955 | * ata_scsi_dev_rescan - initiate scsi_rescan_device() |
3956 | * @work: Pointer to ATA port to perform scsi_rescan_device() |
3957 | * |
3958 | * After ATA pass thru (SAT) commands are executed successfully, |
3959 | * libata need to propagate the changes to SCSI layer. |
3960 | * |
3961 | * LOCKING: |
3962 | * Kernel thread context (may sleep). |
3963 | */ |
3964 | void ata_scsi_dev_rescan(struct work_struct *work) |
3965 | { |
3966 | struct ata_port *ap = |
3967 | container_of(work, struct ata_port, scsi_rescan_task); |
3968 | struct ata_link *link; |
3969 | struct ata_device *dev; |
3970 | unsigned long flags; |
3971 | |
3972 | mutex_lock(&ap->scsi_scan_mutex); |
3973 | spin_lock_irqsave(ap->lock, flags); |
3974 | |
3975 | ata_for_each_link(link, ap, EDGE) { |
3976 | ata_for_each_dev(dev, link, ENABLED) { |
3977 | struct scsi_device *sdev = dev->sdev; |
3978 | |
3979 | if (!sdev) |
3980 | continue; |
3981 | if (scsi_device_get(sdev)) |
3982 | continue; |
3983 | |
3984 | spin_unlock_irqrestore(ap->lock, flags); |
3985 | scsi_rescan_device(&(sdev->sdev_gendev)); |
3986 | scsi_device_put(sdev); |
3987 | spin_lock_irqsave(ap->lock, flags); |
3988 | } |
3989 | } |
3990 | |
3991 | spin_unlock_irqrestore(ap->lock, flags); |
3992 | mutex_unlock(&ap->scsi_scan_mutex); |
3993 | } |
3994 | |
3995 | /** |
3996 | * ata_sas_port_alloc - Allocate port for a SAS attached SATA device |
3997 | * @host: ATA host container for all SAS ports |
3998 | * @port_info: Information from low-level host driver |
3999 | * @shost: SCSI host that the scsi device is attached to |
4000 | * |
4001 | * LOCKING: |
4002 | * PCI/etc. bus probe sem. |
4003 | * |
4004 | * RETURNS: |
4005 | * ata_port pointer on success / NULL on failure. |
4006 | */ |
4007 | |
4008 | struct ata_port *ata_sas_port_alloc(struct ata_host *host, |
4009 | struct ata_port_info *port_info, |
4010 | struct Scsi_Host *shost) |
4011 | { |
4012 | struct ata_port *ap; |
4013 | |
4014 | ap = ata_port_alloc(host); |
4015 | if (!ap) |
4016 | return NULL; |
4017 | |
4018 | ap->port_no = 0; |
4019 | ap->lock = &host->lock; |
4020 | ap->pio_mask = port_info->pio_mask; |
4021 | ap->mwdma_mask = port_info->mwdma_mask; |
4022 | ap->udma_mask = port_info->udma_mask; |
4023 | ap->flags |= port_info->flags; |
4024 | ap->ops = port_info->port_ops; |
4025 | ap->cbl = ATA_CBL_SATA; |
4026 | |
4027 | return ap; |
4028 | } |
4029 | EXPORT_SYMBOL_GPL(ata_sas_port_alloc); |
4030 | |
4031 | /** |
4032 | * ata_sas_port_start - Set port up for dma. |
4033 | * @ap: Port to initialize |
4034 | * |
4035 | * Called just after data structures for each port are |
4036 | * initialized. |
4037 | * |
4038 | * May be used as the port_start() entry in ata_port_operations. |
4039 | * |
4040 | * LOCKING: |
4041 | * Inherited from caller. |
4042 | */ |
4043 | int ata_sas_port_start(struct ata_port *ap) |
4044 | { |
4045 | /* |
4046 | * the port is marked as frozen at allocation time, but if we don't |
4047 | * have new eh, we won't thaw it |
4048 | */ |
4049 | if (!ap->ops->error_handler) |
4050 | ap->pflags &= ~ATA_PFLAG_FROZEN; |
4051 | return 0; |
4052 | } |
4053 | EXPORT_SYMBOL_GPL(ata_sas_port_start); |
4054 | |
4055 | /** |
4056 | * ata_port_stop - Undo ata_sas_port_start() |
4057 | * @ap: Port to shut down |
4058 | * |
4059 | * May be used as the port_stop() entry in ata_port_operations. |
4060 | * |
4061 | * LOCKING: |
4062 | * Inherited from caller. |
4063 | */ |
4064 | |
4065 | void ata_sas_port_stop(struct ata_port *ap) |
4066 | { |
4067 | } |
4068 | EXPORT_SYMBOL_GPL(ata_sas_port_stop); |
4069 | |
4070 | /** |
4071 | * ata_sas_async_probe - simply schedule probing and return |
4072 | * @ap: Port to probe |
4073 | * |
4074 | * For batch scheduling of probe for sas attached ata devices, assumes |
4075 | * the port has already been through ata_sas_port_init() |
4076 | */ |
4077 | void ata_sas_async_probe(struct ata_port *ap) |
4078 | { |
4079 | __ata_port_probe(ap); |
4080 | } |
4081 | EXPORT_SYMBOL_GPL(ata_sas_async_probe); |
4082 | |
4083 | int ata_sas_sync_probe(struct ata_port *ap) |
4084 | { |
4085 | return ata_port_probe(ap); |
4086 | } |
4087 | EXPORT_SYMBOL_GPL(ata_sas_sync_probe); |
4088 | |
4089 | |
4090 | /** |
4091 | * ata_sas_port_init - Initialize a SATA device |
4092 | * @ap: SATA port to initialize |
4093 | * |
4094 | * LOCKING: |
4095 | * PCI/etc. bus probe sem. |
4096 | * |
4097 | * RETURNS: |
4098 | * Zero on success, non-zero on error. |
4099 | */ |
4100 | |
4101 | int ata_sas_port_init(struct ata_port *ap) |
4102 | { |
4103 | int rc = ap->ops->port_start(ap); |
4104 | |
4105 | if (rc) |
4106 | return rc; |
4107 | ap->print_id = atomic_inc_return(&ata_print_id); |
4108 | return 0; |
4109 | } |
4110 | EXPORT_SYMBOL_GPL(ata_sas_port_init); |
4111 | |
4112 | /** |
4113 | * ata_sas_port_destroy - Destroy a SATA port allocated by ata_sas_port_alloc |
4114 | * @ap: SATA port to destroy |
4115 | * |
4116 | */ |
4117 | |
4118 | void ata_sas_port_destroy(struct ata_port *ap) |
4119 | { |
4120 | if (ap->ops->port_stop) |
4121 | ap->ops->port_stop(ap); |
4122 | kfree(ap); |
4123 | } |
4124 | EXPORT_SYMBOL_GPL(ata_sas_port_destroy); |
4125 | |
4126 | /** |
4127 | * ata_sas_slave_configure - Default slave_config routine for libata devices |
4128 | * @sdev: SCSI device to configure |
4129 | * @ap: ATA port to which SCSI device is attached |
4130 | * |
4131 | * RETURNS: |
4132 | * Zero. |
4133 | */ |
4134 | |
4135 | int ata_sas_slave_configure(struct scsi_device *sdev, struct ata_port *ap) |
4136 | { |
4137 | ata_scsi_sdev_config(sdev); |
4138 | ata_scsi_dev_config(sdev, ap->link.device); |
4139 | return 0; |
4140 | } |
4141 | EXPORT_SYMBOL_GPL(ata_sas_slave_configure); |
4142 | |
4143 | /** |
4144 | * ata_sas_queuecmd - Issue SCSI cdb to libata-managed device |
4145 | * @cmd: SCSI command to be sent |
4146 | * @ap: ATA port to which the command is being sent |
4147 | * |
4148 | * RETURNS: |
4149 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, |
4150 | * 0 otherwise. |
4151 | */ |
4152 | |
4153 | int ata_sas_queuecmd(struct scsi_cmnd *cmd, struct ata_port *ap) |
4154 | { |
4155 | int rc = 0; |
4156 | |
4157 | ata_scsi_dump_cdb(ap, cmd); |
4158 | |
4159 | if (likely(ata_dev_enabled(ap->link.device))) |
4160 | rc = __ata_scsi_queuecmd(cmd, ap->link.device); |
4161 | else { |
4162 | cmd->result = (DID_BAD_TARGET << 16); |
4163 | cmd->scsi_done(cmd); |
4164 | } |
4165 | return rc; |
4166 | } |
4167 | EXPORT_SYMBOL_GPL(ata_sas_queuecmd); |
4168 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9