Root/
1 | /* |
2 | * libata-scsi.c - helper library for ATA |
3 | * |
4 | * Maintained by: Jeff Garzik <jgarzik@pobox.com> |
5 | * Please ALWAYS copy linux-ide@vger.kernel.org |
6 | * on emails. |
7 | * |
8 | * Copyright 2003-2004 Red Hat, Inc. All rights reserved. |
9 | * Copyright 2003-2004 Jeff Garzik |
10 | * |
11 | * |
12 | * This program is free software; you can redistribute it and/or modify |
13 | * it under the terms of the GNU General Public License as published by |
14 | * the Free Software Foundation; either version 2, or (at your option) |
15 | * any later version. |
16 | * |
17 | * This program is distributed in the hope that it will be useful, |
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
20 | * GNU General Public License for more details. |
21 | * |
22 | * You should have received a copy of the GNU General Public License |
23 | * along with this program; see the file COPYING. If not, write to |
24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. |
25 | * |
26 | * |
27 | * libata documentation is available via 'make {ps|pdf}docs', |
28 | * as Documentation/DocBook/libata.* |
29 | * |
30 | * Hardware documentation available from |
31 | * - http://www.t10.org/ |
32 | * - http://www.t13.org/ |
33 | * |
34 | */ |
35 | |
36 | #include <linux/slab.h> |
37 | #include <linux/kernel.h> |
38 | #include <linux/blkdev.h> |
39 | #include <linux/spinlock.h> |
40 | #include <linux/export.h> |
41 | #include <scsi/scsi.h> |
42 | #include <scsi/scsi_host.h> |
43 | #include <scsi/scsi_cmnd.h> |
44 | #include <scsi/scsi_eh.h> |
45 | #include <scsi/scsi_device.h> |
46 | #include <scsi/scsi_tcq.h> |
47 | #include <scsi/scsi_transport.h> |
48 | #include <linux/libata.h> |
49 | #include <linux/hdreg.h> |
50 | #include <linux/uaccess.h> |
51 | #include <linux/suspend.h> |
52 | #include <asm/unaligned.h> |
53 | |
54 | #include "libata.h" |
55 | #include "libata-transport.h" |
56 | |
57 | #define ATA_SCSI_RBUF_SIZE 4096 |
58 | |
59 | static DEFINE_SPINLOCK(ata_scsi_rbuf_lock); |
60 | static u8 ata_scsi_rbuf[ATA_SCSI_RBUF_SIZE]; |
61 | |
62 | typedef unsigned int (*ata_xlat_func_t)(struct ata_queued_cmd *qc); |
63 | |
64 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, |
65 | const struct scsi_device *scsidev); |
66 | static struct ata_device *ata_scsi_find_dev(struct ata_port *ap, |
67 | const struct scsi_device *scsidev); |
68 | |
69 | #define RW_RECOVERY_MPAGE 0x1 |
70 | #define RW_RECOVERY_MPAGE_LEN 12 |
71 | #define CACHE_MPAGE 0x8 |
72 | #define CACHE_MPAGE_LEN 20 |
73 | #define CONTROL_MPAGE 0xa |
74 | #define CONTROL_MPAGE_LEN 12 |
75 | #define ALL_MPAGES 0x3f |
76 | #define ALL_SUB_MPAGES 0xff |
77 | |
78 | |
79 | static const u8 def_rw_recovery_mpage[RW_RECOVERY_MPAGE_LEN] = { |
80 | RW_RECOVERY_MPAGE, |
81 | RW_RECOVERY_MPAGE_LEN - 2, |
82 | (1 << 7), /* AWRE */ |
83 | 0, /* read retry count */ |
84 | 0, 0, 0, 0, |
85 | 0, /* write retry count */ |
86 | 0, 0, 0 |
87 | }; |
88 | |
89 | static const u8 def_cache_mpage[CACHE_MPAGE_LEN] = { |
90 | CACHE_MPAGE, |
91 | CACHE_MPAGE_LEN - 2, |
92 | 0, /* contains WCE, needs to be 0 for logic */ |
93 | 0, 0, 0, 0, 0, 0, 0, 0, 0, |
94 | 0, /* contains DRA, needs to be 0 for logic */ |
95 | 0, 0, 0, 0, 0, 0, 0 |
96 | }; |
97 | |
98 | static const u8 def_control_mpage[CONTROL_MPAGE_LEN] = { |
99 | CONTROL_MPAGE, |
100 | CONTROL_MPAGE_LEN - 2, |
101 | 2, /* DSENSE=0, GLTSD=1 */ |
102 | 0, /* [QAM+QERR may be 1, see 05-359r1] */ |
103 | 0, 0, 0, 0, 0xff, 0xff, |
104 | 0, 30 /* extended self test time, see 05-359r1 */ |
105 | }; |
106 | |
107 | static const char *ata_lpm_policy_names[] = { |
108 | [ATA_LPM_UNKNOWN] = "max_performance", |
109 | [ATA_LPM_MAX_POWER] = "max_performance", |
110 | [ATA_LPM_MED_POWER] = "medium_power", |
111 | [ATA_LPM_MIN_POWER] = "min_power", |
112 | }; |
113 | |
114 | static ssize_t ata_scsi_lpm_store(struct device *dev, |
115 | struct device_attribute *attr, |
116 | const char *buf, size_t count) |
117 | { |
118 | struct Scsi_Host *shost = class_to_shost(dev); |
119 | struct ata_port *ap = ata_shost_to_port(shost); |
120 | enum ata_lpm_policy policy; |
121 | unsigned long flags; |
122 | |
123 | /* UNKNOWN is internal state, iterate from MAX_POWER */ |
124 | for (policy = ATA_LPM_MAX_POWER; |
125 | policy < ARRAY_SIZE(ata_lpm_policy_names); policy++) { |
126 | const char *name = ata_lpm_policy_names[policy]; |
127 | |
128 | if (strncmp(name, buf, strlen(name)) == 0) |
129 | break; |
130 | } |
131 | if (policy == ARRAY_SIZE(ata_lpm_policy_names)) |
132 | return -EINVAL; |
133 | |
134 | spin_lock_irqsave(ap->lock, flags); |
135 | ap->target_lpm_policy = policy; |
136 | ata_port_schedule_eh(ap); |
137 | spin_unlock_irqrestore(ap->lock, flags); |
138 | |
139 | return count; |
140 | } |
141 | |
142 | static ssize_t ata_scsi_lpm_show(struct device *dev, |
143 | struct device_attribute *attr, char *buf) |
144 | { |
145 | struct Scsi_Host *shost = class_to_shost(dev); |
146 | struct ata_port *ap = ata_shost_to_port(shost); |
147 | |
148 | if (ap->target_lpm_policy >= ARRAY_SIZE(ata_lpm_policy_names)) |
149 | return -EINVAL; |
150 | |
151 | return snprintf(buf, PAGE_SIZE, "%s\n", |
152 | ata_lpm_policy_names[ap->target_lpm_policy]); |
153 | } |
154 | DEVICE_ATTR(link_power_management_policy, S_IRUGO | S_IWUSR, |
155 | ata_scsi_lpm_show, ata_scsi_lpm_store); |
156 | EXPORT_SYMBOL_GPL(dev_attr_link_power_management_policy); |
157 | |
158 | static ssize_t ata_scsi_park_show(struct device *device, |
159 | struct device_attribute *attr, char *buf) |
160 | { |
161 | struct scsi_device *sdev = to_scsi_device(device); |
162 | struct ata_port *ap; |
163 | struct ata_link *link; |
164 | struct ata_device *dev; |
165 | unsigned long flags, now; |
166 | unsigned int uninitialized_var(msecs); |
167 | int rc = 0; |
168 | |
169 | ap = ata_shost_to_port(sdev->host); |
170 | |
171 | spin_lock_irqsave(ap->lock, flags); |
172 | dev = ata_scsi_find_dev(ap, sdev); |
173 | if (!dev) { |
174 | rc = -ENODEV; |
175 | goto unlock; |
176 | } |
177 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { |
178 | rc = -EOPNOTSUPP; |
179 | goto unlock; |
180 | } |
181 | |
182 | link = dev->link; |
183 | now = jiffies; |
184 | if (ap->pflags & ATA_PFLAG_EH_IN_PROGRESS && |
185 | link->eh_context.unloaded_mask & (1 << dev->devno) && |
186 | time_after(dev->unpark_deadline, now)) |
187 | msecs = jiffies_to_msecs(dev->unpark_deadline - now); |
188 | else |
189 | msecs = 0; |
190 | |
191 | unlock: |
192 | spin_unlock_irq(ap->lock); |
193 | |
194 | return rc ? rc : snprintf(buf, 20, "%u\n", msecs); |
195 | } |
196 | |
197 | static ssize_t ata_scsi_park_store(struct device *device, |
198 | struct device_attribute *attr, |
199 | const char *buf, size_t len) |
200 | { |
201 | struct scsi_device *sdev = to_scsi_device(device); |
202 | struct ata_port *ap; |
203 | struct ata_device *dev; |
204 | long int input; |
205 | unsigned long flags; |
206 | int rc; |
207 | |
208 | rc = strict_strtol(buf, 10, &input); |
209 | if (rc || input < -2) |
210 | return -EINVAL; |
211 | if (input > ATA_TMOUT_MAX_PARK) { |
212 | rc = -EOVERFLOW; |
213 | input = ATA_TMOUT_MAX_PARK; |
214 | } |
215 | |
216 | ap = ata_shost_to_port(sdev->host); |
217 | |
218 | spin_lock_irqsave(ap->lock, flags); |
219 | dev = ata_scsi_find_dev(ap, sdev); |
220 | if (unlikely(!dev)) { |
221 | rc = -ENODEV; |
222 | goto unlock; |
223 | } |
224 | if (dev->class != ATA_DEV_ATA) { |
225 | rc = -EOPNOTSUPP; |
226 | goto unlock; |
227 | } |
228 | |
229 | if (input >= 0) { |
230 | if (dev->flags & ATA_DFLAG_NO_UNLOAD) { |
231 | rc = -EOPNOTSUPP; |
232 | goto unlock; |
233 | } |
234 | |
235 | dev->unpark_deadline = ata_deadline(jiffies, input); |
236 | dev->link->eh_info.dev_action[dev->devno] |= ATA_EH_PARK; |
237 | ata_port_schedule_eh(ap); |
238 | complete(&ap->park_req_pending); |
239 | } else { |
240 | switch (input) { |
241 | case -1: |
242 | dev->flags &= ~ATA_DFLAG_NO_UNLOAD; |
243 | break; |
244 | case -2: |
245 | dev->flags |= ATA_DFLAG_NO_UNLOAD; |
246 | break; |
247 | } |
248 | } |
249 | unlock: |
250 | spin_unlock_irqrestore(ap->lock, flags); |
251 | |
252 | return rc ? rc : len; |
253 | } |
254 | DEVICE_ATTR(unload_heads, S_IRUGO | S_IWUSR, |
255 | ata_scsi_park_show, ata_scsi_park_store); |
256 | EXPORT_SYMBOL_GPL(dev_attr_unload_heads); |
257 | |
258 | static void ata_scsi_set_sense(struct scsi_cmnd *cmd, u8 sk, u8 asc, u8 ascq) |
259 | { |
260 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
261 | |
262 | scsi_build_sense_buffer(0, cmd->sense_buffer, sk, asc, ascq); |
263 | } |
264 | |
265 | static ssize_t |
266 | ata_scsi_em_message_store(struct device *dev, struct device_attribute *attr, |
267 | const char *buf, size_t count) |
268 | { |
269 | struct Scsi_Host *shost = class_to_shost(dev); |
270 | struct ata_port *ap = ata_shost_to_port(shost); |
271 | if (ap->ops->em_store && (ap->flags & ATA_FLAG_EM)) |
272 | return ap->ops->em_store(ap, buf, count); |
273 | return -EINVAL; |
274 | } |
275 | |
276 | static ssize_t |
277 | ata_scsi_em_message_show(struct device *dev, struct device_attribute *attr, |
278 | char *buf) |
279 | { |
280 | struct Scsi_Host *shost = class_to_shost(dev); |
281 | struct ata_port *ap = ata_shost_to_port(shost); |
282 | |
283 | if (ap->ops->em_show && (ap->flags & ATA_FLAG_EM)) |
284 | return ap->ops->em_show(ap, buf); |
285 | return -EINVAL; |
286 | } |
287 | DEVICE_ATTR(em_message, S_IRUGO | S_IWUSR, |
288 | ata_scsi_em_message_show, ata_scsi_em_message_store); |
289 | EXPORT_SYMBOL_GPL(dev_attr_em_message); |
290 | |
291 | static ssize_t |
292 | ata_scsi_em_message_type_show(struct device *dev, struct device_attribute *attr, |
293 | char *buf) |
294 | { |
295 | struct Scsi_Host *shost = class_to_shost(dev); |
296 | struct ata_port *ap = ata_shost_to_port(shost); |
297 | |
298 | return snprintf(buf, 23, "%d\n", ap->em_message_type); |
299 | } |
300 | DEVICE_ATTR(em_message_type, S_IRUGO, |
301 | ata_scsi_em_message_type_show, NULL); |
302 | EXPORT_SYMBOL_GPL(dev_attr_em_message_type); |
303 | |
304 | static ssize_t |
305 | ata_scsi_activity_show(struct device *dev, struct device_attribute *attr, |
306 | char *buf) |
307 | { |
308 | struct scsi_device *sdev = to_scsi_device(dev); |
309 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
310 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); |
311 | |
312 | if (ap->ops->sw_activity_show && (ap->flags & ATA_FLAG_SW_ACTIVITY)) |
313 | return ap->ops->sw_activity_show(atadev, buf); |
314 | return -EINVAL; |
315 | } |
316 | |
317 | static ssize_t |
318 | ata_scsi_activity_store(struct device *dev, struct device_attribute *attr, |
319 | const char *buf, size_t count) |
320 | { |
321 | struct scsi_device *sdev = to_scsi_device(dev); |
322 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
323 | struct ata_device *atadev = ata_scsi_find_dev(ap, sdev); |
324 | enum sw_activity val; |
325 | int rc; |
326 | |
327 | if (ap->ops->sw_activity_store && (ap->flags & ATA_FLAG_SW_ACTIVITY)) { |
328 | val = simple_strtoul(buf, NULL, 0); |
329 | switch (val) { |
330 | case OFF: case BLINK_ON: case BLINK_OFF: |
331 | rc = ap->ops->sw_activity_store(atadev, val); |
332 | if (!rc) |
333 | return count; |
334 | else |
335 | return rc; |
336 | } |
337 | } |
338 | return -EINVAL; |
339 | } |
340 | DEVICE_ATTR(sw_activity, S_IWUSR | S_IRUGO, ata_scsi_activity_show, |
341 | ata_scsi_activity_store); |
342 | EXPORT_SYMBOL_GPL(dev_attr_sw_activity); |
343 | |
344 | struct device_attribute *ata_common_sdev_attrs[] = { |
345 | &dev_attr_unload_heads, |
346 | NULL |
347 | }; |
348 | EXPORT_SYMBOL_GPL(ata_common_sdev_attrs); |
349 | |
350 | static void ata_scsi_invalid_field(struct scsi_cmnd *cmd) |
351 | { |
352 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x24, 0x0); |
353 | /* "Invalid field in cbd" */ |
354 | cmd->scsi_done(cmd); |
355 | } |
356 | |
357 | /** |
358 | * ata_std_bios_param - generic bios head/sector/cylinder calculator used by sd. |
359 | * @sdev: SCSI device for which BIOS geometry is to be determined |
360 | * @bdev: block device associated with @sdev |
361 | * @capacity: capacity of SCSI device |
362 | * @geom: location to which geometry will be output |
363 | * |
364 | * Generic bios head/sector/cylinder calculator |
365 | * used by sd. Most BIOSes nowadays expect a XXX/255/16 (CHS) |
366 | * mapping. Some situations may arise where the disk is not |
367 | * bootable if this is not used. |
368 | * |
369 | * LOCKING: |
370 | * Defined by the SCSI layer. We don't really care. |
371 | * |
372 | * RETURNS: |
373 | * Zero. |
374 | */ |
375 | int ata_std_bios_param(struct scsi_device *sdev, struct block_device *bdev, |
376 | sector_t capacity, int geom[]) |
377 | { |
378 | geom[0] = 255; |
379 | geom[1] = 63; |
380 | sector_div(capacity, 255*63); |
381 | geom[2] = capacity; |
382 | |
383 | return 0; |
384 | } |
385 | |
386 | /** |
387 | * ata_scsi_unlock_native_capacity - unlock native capacity |
388 | * @sdev: SCSI device to adjust device capacity for |
389 | * |
390 | * This function is called if a partition on @sdev extends beyond |
391 | * the end of the device. It requests EH to unlock HPA. |
392 | * |
393 | * LOCKING: |
394 | * Defined by the SCSI layer. Might sleep. |
395 | */ |
396 | void ata_scsi_unlock_native_capacity(struct scsi_device *sdev) |
397 | { |
398 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
399 | struct ata_device *dev; |
400 | unsigned long flags; |
401 | |
402 | spin_lock_irqsave(ap->lock, flags); |
403 | |
404 | dev = ata_scsi_find_dev(ap, sdev); |
405 | if (dev && dev->n_sectors < dev->n_native_sectors) { |
406 | dev->flags |= ATA_DFLAG_UNLOCK_HPA; |
407 | dev->link->eh_info.action |= ATA_EH_RESET; |
408 | ata_port_schedule_eh(ap); |
409 | } |
410 | |
411 | spin_unlock_irqrestore(ap->lock, flags); |
412 | ata_port_wait_eh(ap); |
413 | } |
414 | |
415 | /** |
416 | * ata_get_identity - Handler for HDIO_GET_IDENTITY ioctl |
417 | * @ap: target port |
418 | * @sdev: SCSI device to get identify data for |
419 | * @arg: User buffer area for identify data |
420 | * |
421 | * LOCKING: |
422 | * Defined by the SCSI layer. We don't really care. |
423 | * |
424 | * RETURNS: |
425 | * Zero on success, negative errno on error. |
426 | */ |
427 | static int ata_get_identity(struct ata_port *ap, struct scsi_device *sdev, |
428 | void __user *arg) |
429 | { |
430 | struct ata_device *dev = ata_scsi_find_dev(ap, sdev); |
431 | u16 __user *dst = arg; |
432 | char buf[40]; |
433 | |
434 | if (!dev) |
435 | return -ENOMSG; |
436 | |
437 | if (copy_to_user(dst, dev->id, ATA_ID_WORDS * sizeof(u16))) |
438 | return -EFAULT; |
439 | |
440 | ata_id_string(dev->id, buf, ATA_ID_PROD, ATA_ID_PROD_LEN); |
441 | if (copy_to_user(dst + ATA_ID_PROD, buf, ATA_ID_PROD_LEN)) |
442 | return -EFAULT; |
443 | |
444 | ata_id_string(dev->id, buf, ATA_ID_FW_REV, ATA_ID_FW_REV_LEN); |
445 | if (copy_to_user(dst + ATA_ID_FW_REV, buf, ATA_ID_FW_REV_LEN)) |
446 | return -EFAULT; |
447 | |
448 | ata_id_string(dev->id, buf, ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
449 | if (copy_to_user(dst + ATA_ID_SERNO, buf, ATA_ID_SERNO_LEN)) |
450 | return -EFAULT; |
451 | |
452 | return 0; |
453 | } |
454 | |
455 | /** |
456 | * ata_cmd_ioctl - Handler for HDIO_DRIVE_CMD ioctl |
457 | * @scsidev: Device to which we are issuing command |
458 | * @arg: User provided data for issuing command |
459 | * |
460 | * LOCKING: |
461 | * Defined by the SCSI layer. We don't really care. |
462 | * |
463 | * RETURNS: |
464 | * Zero on success, negative errno on error. |
465 | */ |
466 | int ata_cmd_ioctl(struct scsi_device *scsidev, void __user *arg) |
467 | { |
468 | int rc = 0; |
469 | u8 scsi_cmd[MAX_COMMAND_SIZE]; |
470 | u8 args[4], *argbuf = NULL, *sensebuf = NULL; |
471 | int argsize = 0; |
472 | enum dma_data_direction data_dir; |
473 | int cmd_result; |
474 | |
475 | if (arg == NULL) |
476 | return -EINVAL; |
477 | |
478 | if (copy_from_user(args, arg, sizeof(args))) |
479 | return -EFAULT; |
480 | |
481 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); |
482 | if (!sensebuf) |
483 | return -ENOMEM; |
484 | |
485 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); |
486 | |
487 | if (args[3]) { |
488 | argsize = ATA_SECT_SIZE * args[3]; |
489 | argbuf = kmalloc(argsize, GFP_KERNEL); |
490 | if (argbuf == NULL) { |
491 | rc = -ENOMEM; |
492 | goto error; |
493 | } |
494 | |
495 | scsi_cmd[1] = (4 << 1); /* PIO Data-in */ |
496 | scsi_cmd[2] = 0x0e; /* no off.line or cc, read from dev, |
497 | block count in sector count field */ |
498 | data_dir = DMA_FROM_DEVICE; |
499 | } else { |
500 | scsi_cmd[1] = (3 << 1); /* Non-data */ |
501 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ |
502 | data_dir = DMA_NONE; |
503 | } |
504 | |
505 | scsi_cmd[0] = ATA_16; |
506 | |
507 | scsi_cmd[4] = args[2]; |
508 | if (args[0] == ATA_CMD_SMART) { /* hack -- ide driver does this too */ |
509 | scsi_cmd[6] = args[3]; |
510 | scsi_cmd[8] = args[1]; |
511 | scsi_cmd[10] = 0x4f; |
512 | scsi_cmd[12] = 0xc2; |
513 | } else { |
514 | scsi_cmd[6] = args[1]; |
515 | } |
516 | scsi_cmd[14] = args[0]; |
517 | |
518 | /* Good values for timeout and retries? Values below |
519 | from scsi_ioctl_send_command() for default case... */ |
520 | cmd_result = scsi_execute(scsidev, scsi_cmd, data_dir, argbuf, argsize, |
521 | sensebuf, (10*HZ), 5, 0, NULL); |
522 | |
523 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ |
524 | u8 *desc = sensebuf + 8; |
525 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ |
526 | |
527 | /* If we set cc then ATA pass-through will cause a |
528 | * check condition even if no error. Filter that. */ |
529 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { |
530 | struct scsi_sense_hdr sshdr; |
531 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, |
532 | &sshdr); |
533 | if (sshdr.sense_key == 0 && |
534 | sshdr.asc == 0 && sshdr.ascq == 0) |
535 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; |
536 | } |
537 | |
538 | /* Send userspace a few ATA registers (same as drivers/ide) */ |
539 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ |
540 | desc[0] == 0x09) { /* code is "ATA Descriptor" */ |
541 | args[0] = desc[13]; /* status */ |
542 | args[1] = desc[3]; /* error */ |
543 | args[2] = desc[5]; /* sector count (0:7) */ |
544 | if (copy_to_user(arg, args, sizeof(args))) |
545 | rc = -EFAULT; |
546 | } |
547 | } |
548 | |
549 | |
550 | if (cmd_result) { |
551 | rc = -EIO; |
552 | goto error; |
553 | } |
554 | |
555 | if ((argbuf) |
556 | && copy_to_user(arg + sizeof(args), argbuf, argsize)) |
557 | rc = -EFAULT; |
558 | error: |
559 | kfree(sensebuf); |
560 | kfree(argbuf); |
561 | return rc; |
562 | } |
563 | |
564 | /** |
565 | * ata_task_ioctl - Handler for HDIO_DRIVE_TASK ioctl |
566 | * @scsidev: Device to which we are issuing command |
567 | * @arg: User provided data for issuing command |
568 | * |
569 | * LOCKING: |
570 | * Defined by the SCSI layer. We don't really care. |
571 | * |
572 | * RETURNS: |
573 | * Zero on success, negative errno on error. |
574 | */ |
575 | int ata_task_ioctl(struct scsi_device *scsidev, void __user *arg) |
576 | { |
577 | int rc = 0; |
578 | u8 scsi_cmd[MAX_COMMAND_SIZE]; |
579 | u8 args[7], *sensebuf = NULL; |
580 | int cmd_result; |
581 | |
582 | if (arg == NULL) |
583 | return -EINVAL; |
584 | |
585 | if (copy_from_user(args, arg, sizeof(args))) |
586 | return -EFAULT; |
587 | |
588 | sensebuf = kzalloc(SCSI_SENSE_BUFFERSIZE, GFP_NOIO); |
589 | if (!sensebuf) |
590 | return -ENOMEM; |
591 | |
592 | memset(scsi_cmd, 0, sizeof(scsi_cmd)); |
593 | scsi_cmd[0] = ATA_16; |
594 | scsi_cmd[1] = (3 << 1); /* Non-data */ |
595 | scsi_cmd[2] = 0x20; /* cc but no off.line or data xfer */ |
596 | scsi_cmd[4] = args[1]; |
597 | scsi_cmd[6] = args[2]; |
598 | scsi_cmd[8] = args[3]; |
599 | scsi_cmd[10] = args[4]; |
600 | scsi_cmd[12] = args[5]; |
601 | scsi_cmd[13] = args[6] & 0x4f; |
602 | scsi_cmd[14] = args[0]; |
603 | |
604 | /* Good values for timeout and retries? Values below |
605 | from scsi_ioctl_send_command() for default case... */ |
606 | cmd_result = scsi_execute(scsidev, scsi_cmd, DMA_NONE, NULL, 0, |
607 | sensebuf, (10*HZ), 5, 0, NULL); |
608 | |
609 | if (driver_byte(cmd_result) == DRIVER_SENSE) {/* sense data available */ |
610 | u8 *desc = sensebuf + 8; |
611 | cmd_result &= ~(0xFF<<24); /* DRIVER_SENSE is not an error */ |
612 | |
613 | /* If we set cc then ATA pass-through will cause a |
614 | * check condition even if no error. Filter that. */ |
615 | if (cmd_result & SAM_STAT_CHECK_CONDITION) { |
616 | struct scsi_sense_hdr sshdr; |
617 | scsi_normalize_sense(sensebuf, SCSI_SENSE_BUFFERSIZE, |
618 | &sshdr); |
619 | if (sshdr.sense_key == 0 && |
620 | sshdr.asc == 0 && sshdr.ascq == 0) |
621 | cmd_result &= ~SAM_STAT_CHECK_CONDITION; |
622 | } |
623 | |
624 | /* Send userspace ATA registers */ |
625 | if (sensebuf[0] == 0x72 && /* format is "descriptor" */ |
626 | desc[0] == 0x09) {/* code is "ATA Descriptor" */ |
627 | args[0] = desc[13]; /* status */ |
628 | args[1] = desc[3]; /* error */ |
629 | args[2] = desc[5]; /* sector count (0:7) */ |
630 | args[3] = desc[7]; /* lbal */ |
631 | args[4] = desc[9]; /* lbam */ |
632 | args[5] = desc[11]; /* lbah */ |
633 | args[6] = desc[12]; /* select */ |
634 | if (copy_to_user(arg, args, sizeof(args))) |
635 | rc = -EFAULT; |
636 | } |
637 | } |
638 | |
639 | if (cmd_result) { |
640 | rc = -EIO; |
641 | goto error; |
642 | } |
643 | |
644 | error: |
645 | kfree(sensebuf); |
646 | return rc; |
647 | } |
648 | |
649 | static int ata_ioc32(struct ata_port *ap) |
650 | { |
651 | if (ap->flags & ATA_FLAG_PIO_DMA) |
652 | return 1; |
653 | if (ap->pflags & ATA_PFLAG_PIO32) |
654 | return 1; |
655 | return 0; |
656 | } |
657 | |
658 | int ata_sas_scsi_ioctl(struct ata_port *ap, struct scsi_device *scsidev, |
659 | int cmd, void __user *arg) |
660 | { |
661 | int val = -EINVAL, rc = -EINVAL; |
662 | unsigned long flags; |
663 | |
664 | switch (cmd) { |
665 | case ATA_IOC_GET_IO32: |
666 | spin_lock_irqsave(ap->lock, flags); |
667 | val = ata_ioc32(ap); |
668 | spin_unlock_irqrestore(ap->lock, flags); |
669 | if (copy_to_user(arg, &val, 1)) |
670 | return -EFAULT; |
671 | return 0; |
672 | |
673 | case ATA_IOC_SET_IO32: |
674 | val = (unsigned long) arg; |
675 | rc = 0; |
676 | spin_lock_irqsave(ap->lock, flags); |
677 | if (ap->pflags & ATA_PFLAG_PIO32CHANGE) { |
678 | if (val) |
679 | ap->pflags |= ATA_PFLAG_PIO32; |
680 | else |
681 | ap->pflags &= ~ATA_PFLAG_PIO32; |
682 | } else { |
683 | if (val != ata_ioc32(ap)) |
684 | rc = -EINVAL; |
685 | } |
686 | spin_unlock_irqrestore(ap->lock, flags); |
687 | return rc; |
688 | |
689 | case HDIO_GET_IDENTITY: |
690 | return ata_get_identity(ap, scsidev, arg); |
691 | |
692 | case HDIO_DRIVE_CMD: |
693 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) |
694 | return -EACCES; |
695 | return ata_cmd_ioctl(scsidev, arg); |
696 | |
697 | case HDIO_DRIVE_TASK: |
698 | if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO)) |
699 | return -EACCES; |
700 | return ata_task_ioctl(scsidev, arg); |
701 | |
702 | default: |
703 | rc = -ENOTTY; |
704 | break; |
705 | } |
706 | |
707 | return rc; |
708 | } |
709 | EXPORT_SYMBOL_GPL(ata_sas_scsi_ioctl); |
710 | |
711 | int ata_scsi_ioctl(struct scsi_device *scsidev, int cmd, void __user *arg) |
712 | { |
713 | return ata_sas_scsi_ioctl(ata_shost_to_port(scsidev->host), |
714 | scsidev, cmd, arg); |
715 | } |
716 | EXPORT_SYMBOL_GPL(ata_scsi_ioctl); |
717 | |
718 | /** |
719 | * ata_scsi_qc_new - acquire new ata_queued_cmd reference |
720 | * @dev: ATA device to which the new command is attached |
721 | * @cmd: SCSI command that originated this ATA command |
722 | * |
723 | * Obtain a reference to an unused ata_queued_cmd structure, |
724 | * which is the basic libata structure representing a single |
725 | * ATA command sent to the hardware. |
726 | * |
727 | * If a command was available, fill in the SCSI-specific |
728 | * portions of the structure with information on the |
729 | * current command. |
730 | * |
731 | * LOCKING: |
732 | * spin_lock_irqsave(host lock) |
733 | * |
734 | * RETURNS: |
735 | * Command allocated, or %NULL if none available. |
736 | */ |
737 | static struct ata_queued_cmd *ata_scsi_qc_new(struct ata_device *dev, |
738 | struct scsi_cmnd *cmd) |
739 | { |
740 | struct ata_queued_cmd *qc; |
741 | |
742 | qc = ata_qc_new_init(dev); |
743 | if (qc) { |
744 | qc->scsicmd = cmd; |
745 | qc->scsidone = cmd->scsi_done; |
746 | |
747 | qc->sg = scsi_sglist(cmd); |
748 | qc->n_elem = scsi_sg_count(cmd); |
749 | } else { |
750 | cmd->result = (DID_OK << 16) | (QUEUE_FULL << 1); |
751 | cmd->scsi_done(cmd); |
752 | } |
753 | |
754 | return qc; |
755 | } |
756 | |
757 | static void ata_qc_set_pc_nbytes(struct ata_queued_cmd *qc) |
758 | { |
759 | struct scsi_cmnd *scmd = qc->scsicmd; |
760 | |
761 | qc->extrabytes = scmd->request->extra_len; |
762 | qc->nbytes = scsi_bufflen(scmd) + qc->extrabytes; |
763 | } |
764 | |
765 | /** |
766 | * ata_dump_status - user friendly display of error info |
767 | * @id: id of the port in question |
768 | * @tf: ptr to filled out taskfile |
769 | * |
770 | * Decode and dump the ATA error/status registers for the user so |
771 | * that they have some idea what really happened at the non |
772 | * make-believe layer. |
773 | * |
774 | * LOCKING: |
775 | * inherited from caller |
776 | */ |
777 | static void ata_dump_status(unsigned id, struct ata_taskfile *tf) |
778 | { |
779 | u8 stat = tf->command, err = tf->feature; |
780 | |
781 | printk(KERN_WARNING "ata%u: status=0x%02x { ", id, stat); |
782 | if (stat & ATA_BUSY) { |
783 | printk("Busy }\n"); /* Data is not valid in this case */ |
784 | } else { |
785 | if (stat & 0x40) printk("DriveReady "); |
786 | if (stat & 0x20) printk("DeviceFault "); |
787 | if (stat & 0x10) printk("SeekComplete "); |
788 | if (stat & 0x08) printk("DataRequest "); |
789 | if (stat & 0x04) printk("CorrectedError "); |
790 | if (stat & 0x02) printk("Index "); |
791 | if (stat & 0x01) printk("Error "); |
792 | printk("}\n"); |
793 | |
794 | if (err) { |
795 | printk(KERN_WARNING "ata%u: error=0x%02x { ", id, err); |
796 | if (err & 0x04) printk("DriveStatusError "); |
797 | if (err & 0x80) { |
798 | if (err & 0x04) printk("BadCRC "); |
799 | else printk("Sector "); |
800 | } |
801 | if (err & 0x40) printk("UncorrectableError "); |
802 | if (err & 0x10) printk("SectorIdNotFound "); |
803 | if (err & 0x02) printk("TrackZeroNotFound "); |
804 | if (err & 0x01) printk("AddrMarkNotFound "); |
805 | printk("}\n"); |
806 | } |
807 | } |
808 | } |
809 | |
810 | /** |
811 | * ata_to_sense_error - convert ATA error to SCSI error |
812 | * @id: ATA device number |
813 | * @drv_stat: value contained in ATA status register |
814 | * @drv_err: value contained in ATA error register |
815 | * @sk: the sense key we'll fill out |
816 | * @asc: the additional sense code we'll fill out |
817 | * @ascq: the additional sense code qualifier we'll fill out |
818 | * @verbose: be verbose |
819 | * |
820 | * Converts an ATA error into a SCSI error. Fill out pointers to |
821 | * SK, ASC, and ASCQ bytes for later use in fixed or descriptor |
822 | * format sense blocks. |
823 | * |
824 | * LOCKING: |
825 | * spin_lock_irqsave(host lock) |
826 | */ |
827 | static void ata_to_sense_error(unsigned id, u8 drv_stat, u8 drv_err, u8 *sk, |
828 | u8 *asc, u8 *ascq, int verbose) |
829 | { |
830 | int i; |
831 | |
832 | /* Based on the 3ware driver translation table */ |
833 | static const unsigned char sense_table[][4] = { |
834 | /* BBD|ECC|ID|MAR */ |
835 | {0xd1, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command |
836 | /* BBD|ECC|ID */ |
837 | {0xd0, ABORTED_COMMAND, 0x00, 0x00}, // Device busy Aborted command |
838 | /* ECC|MC|MARK */ |
839 | {0x61, HARDWARE_ERROR, 0x00, 0x00}, // Device fault Hardware error |
840 | /* ICRC|ABRT */ /* NB: ICRC & !ABRT is BBD */ |
841 | {0x84, ABORTED_COMMAND, 0x47, 0x00}, // Data CRC error SCSI parity error |
842 | /* MC|ID|ABRT|TRK0|MARK */ |
843 | {0x37, NOT_READY, 0x04, 0x00}, // Unit offline Not ready |
844 | /* MCR|MARK */ |
845 | {0x09, NOT_READY, 0x04, 0x00}, // Unrecovered disk error Not ready |
846 | /* Bad address mark */ |
847 | {0x01, MEDIUM_ERROR, 0x13, 0x00}, // Address mark not found Address mark not found for data field |
848 | /* TRK0 */ |
849 | {0x02, HARDWARE_ERROR, 0x00, 0x00}, // Track 0 not found Hardware error |
850 | /* Abort & !ICRC */ |
851 | {0x04, ABORTED_COMMAND, 0x00, 0x00}, // Aborted command Aborted command |
852 | /* Media change request */ |
853 | {0x08, NOT_READY, 0x04, 0x00}, // Media change request FIXME: faking offline |
854 | /* SRV */ |
855 | {0x10, ABORTED_COMMAND, 0x14, 0x00}, // ID not found Recorded entity not found |
856 | /* Media change */ |
857 | {0x08, NOT_READY, 0x04, 0x00}, // Media change FIXME: faking offline |
858 | /* ECC */ |
859 | {0x40, MEDIUM_ERROR, 0x11, 0x04}, // Uncorrectable ECC error Unrecovered read error |
860 | /* BBD - block marked bad */ |
861 | {0x80, MEDIUM_ERROR, 0x11, 0x04}, // Block marked bad Medium error, unrecovered read error |
862 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark |
863 | }; |
864 | static const unsigned char stat_table[][4] = { |
865 | /* Must be first because BUSY means no other bits valid */ |
866 | {0x80, ABORTED_COMMAND, 0x47, 0x00}, // Busy, fake parity for now |
867 | {0x20, HARDWARE_ERROR, 0x00, 0x00}, // Device fault |
868 | {0x08, ABORTED_COMMAND, 0x47, 0x00}, // Timed out in xfer, fake parity for now |
869 | {0x04, RECOVERED_ERROR, 0x11, 0x00}, // Recovered ECC error Medium error, recovered |
870 | {0xFF, 0xFF, 0xFF, 0xFF}, // END mark |
871 | }; |
872 | |
873 | /* |
874 | * Is this an error we can process/parse |
875 | */ |
876 | if (drv_stat & ATA_BUSY) { |
877 | drv_err = 0; /* Ignore the err bits, they're invalid */ |
878 | } |
879 | |
880 | if (drv_err) { |
881 | /* Look for drv_err */ |
882 | for (i = 0; sense_table[i][0] != 0xFF; i++) { |
883 | /* Look for best matches first */ |
884 | if ((sense_table[i][0] & drv_err) == |
885 | sense_table[i][0]) { |
886 | *sk = sense_table[i][1]; |
887 | *asc = sense_table[i][2]; |
888 | *ascq = sense_table[i][3]; |
889 | goto translate_done; |
890 | } |
891 | } |
892 | /* No immediate match */ |
893 | if (verbose) |
894 | printk(KERN_WARNING "ata%u: no sense translation for " |
895 | "error 0x%02x\n", id, drv_err); |
896 | } |
897 | |
898 | /* Fall back to interpreting status bits */ |
899 | for (i = 0; stat_table[i][0] != 0xFF; i++) { |
900 | if (stat_table[i][0] & drv_stat) { |
901 | *sk = stat_table[i][1]; |
902 | *asc = stat_table[i][2]; |
903 | *ascq = stat_table[i][3]; |
904 | goto translate_done; |
905 | } |
906 | } |
907 | /* No error? Undecoded? */ |
908 | if (verbose) |
909 | printk(KERN_WARNING "ata%u: no sense translation for " |
910 | "status: 0x%02x\n", id, drv_stat); |
911 | |
912 | /* We need a sensible error return here, which is tricky, and one |
913 | that won't cause people to do things like return a disk wrongly */ |
914 | *sk = ABORTED_COMMAND; |
915 | *asc = 0x00; |
916 | *ascq = 0x00; |
917 | |
918 | translate_done: |
919 | if (verbose) |
920 | printk(KERN_ERR "ata%u: translated ATA stat/err 0x%02x/%02x " |
921 | "to SCSI SK/ASC/ASCQ 0x%x/%02x/%02x\n", |
922 | id, drv_stat, drv_err, *sk, *asc, *ascq); |
923 | return; |
924 | } |
925 | |
926 | /* |
927 | * ata_gen_passthru_sense - Generate check condition sense block. |
928 | * @qc: Command that completed. |
929 | * |
930 | * This function is specific to the ATA descriptor format sense |
931 | * block specified for the ATA pass through commands. Regardless |
932 | * of whether the command errored or not, return a sense |
933 | * block. Copy all controller registers into the sense |
934 | * block. Clear sense key, ASC & ASCQ if there is no error. |
935 | * |
936 | * LOCKING: |
937 | * None. |
938 | */ |
939 | static void ata_gen_passthru_sense(struct ata_queued_cmd *qc) |
940 | { |
941 | struct scsi_cmnd *cmd = qc->scsicmd; |
942 | struct ata_taskfile *tf = &qc->result_tf; |
943 | unsigned char *sb = cmd->sense_buffer; |
944 | unsigned char *desc = sb + 8; |
945 | int verbose = qc->ap->ops->error_handler == NULL; |
946 | |
947 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); |
948 | |
949 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
950 | |
951 | /* |
952 | * Use ata_to_sense_error() to map status register bits |
953 | * onto sense key, asc & ascq. |
954 | */ |
955 | if (qc->err_mask || |
956 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { |
957 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, |
958 | &sb[1], &sb[2], &sb[3], verbose); |
959 | sb[1] &= 0x0f; |
960 | } |
961 | |
962 | /* |
963 | * Sense data is current and format is descriptor. |
964 | */ |
965 | sb[0] = 0x72; |
966 | |
967 | desc[0] = 0x09; |
968 | |
969 | /* set length of additional sense data */ |
970 | sb[7] = 14; |
971 | desc[1] = 12; |
972 | |
973 | /* |
974 | * Copy registers into sense buffer. |
975 | */ |
976 | desc[2] = 0x00; |
977 | desc[3] = tf->feature; /* == error reg */ |
978 | desc[5] = tf->nsect; |
979 | desc[7] = tf->lbal; |
980 | desc[9] = tf->lbam; |
981 | desc[11] = tf->lbah; |
982 | desc[12] = tf->device; |
983 | desc[13] = tf->command; /* == status reg */ |
984 | |
985 | /* |
986 | * Fill in Extend bit, and the high order bytes |
987 | * if applicable. |
988 | */ |
989 | if (tf->flags & ATA_TFLAG_LBA48) { |
990 | desc[2] |= 0x01; |
991 | desc[4] = tf->hob_nsect; |
992 | desc[6] = tf->hob_lbal; |
993 | desc[8] = tf->hob_lbam; |
994 | desc[10] = tf->hob_lbah; |
995 | } |
996 | } |
997 | |
998 | /** |
999 | * ata_gen_ata_sense - generate a SCSI fixed sense block |
1000 | * @qc: Command that we are erroring out |
1001 | * |
1002 | * Generate sense block for a failed ATA command @qc. Descriptor |
1003 | * format is used to accommodate LBA48 block address. |
1004 | * |
1005 | * LOCKING: |
1006 | * None. |
1007 | */ |
1008 | static void ata_gen_ata_sense(struct ata_queued_cmd *qc) |
1009 | { |
1010 | struct ata_device *dev = qc->dev; |
1011 | struct scsi_cmnd *cmd = qc->scsicmd; |
1012 | struct ata_taskfile *tf = &qc->result_tf; |
1013 | unsigned char *sb = cmd->sense_buffer; |
1014 | unsigned char *desc = sb + 8; |
1015 | int verbose = qc->ap->ops->error_handler == NULL; |
1016 | u64 block; |
1017 | |
1018 | memset(sb, 0, SCSI_SENSE_BUFFERSIZE); |
1019 | |
1020 | cmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; |
1021 | |
1022 | /* sense data is current and format is descriptor */ |
1023 | sb[0] = 0x72; |
1024 | |
1025 | /* Use ata_to_sense_error() to map status register bits |
1026 | * onto sense key, asc & ascq. |
1027 | */ |
1028 | if (qc->err_mask || |
1029 | tf->command & (ATA_BUSY | ATA_DF | ATA_ERR | ATA_DRQ)) { |
1030 | ata_to_sense_error(qc->ap->print_id, tf->command, tf->feature, |
1031 | &sb[1], &sb[2], &sb[3], verbose); |
1032 | sb[1] &= 0x0f; |
1033 | } |
1034 | |
1035 | block = ata_tf_read_block(&qc->result_tf, dev); |
1036 | |
1037 | /* information sense data descriptor */ |
1038 | sb[7] = 12; |
1039 | desc[0] = 0x00; |
1040 | desc[1] = 10; |
1041 | |
1042 | desc[2] |= 0x80; /* valid */ |
1043 | desc[6] = block >> 40; |
1044 | desc[7] = block >> 32; |
1045 | desc[8] = block >> 24; |
1046 | desc[9] = block >> 16; |
1047 | desc[10] = block >> 8; |
1048 | desc[11] = block; |
1049 | } |
1050 | |
1051 | static void ata_scsi_sdev_config(struct scsi_device *sdev) |
1052 | { |
1053 | sdev->use_10_for_rw = 1; |
1054 | sdev->use_10_for_ms = 1; |
1055 | |
1056 | /* Schedule policy is determined by ->qc_defer() callback and |
1057 | * it needs to see every deferred qc. Set dev_blocked to 1 to |
1058 | * prevent SCSI midlayer from automatically deferring |
1059 | * requests. |
1060 | */ |
1061 | sdev->max_device_blocked = 1; |
1062 | } |
1063 | |
1064 | /** |
1065 | * atapi_drain_needed - Check whether data transfer may overflow |
1066 | * @rq: request to be checked |
1067 | * |
1068 | * ATAPI commands which transfer variable length data to host |
1069 | * might overflow due to application error or hardare bug. This |
1070 | * function checks whether overflow should be drained and ignored |
1071 | * for @request. |
1072 | * |
1073 | * LOCKING: |
1074 | * None. |
1075 | * |
1076 | * RETURNS: |
1077 | * 1 if ; otherwise, 0. |
1078 | */ |
1079 | static int atapi_drain_needed(struct request *rq) |
1080 | { |
1081 | if (likely(rq->cmd_type != REQ_TYPE_BLOCK_PC)) |
1082 | return 0; |
1083 | |
1084 | if (!blk_rq_bytes(rq) || (rq->cmd_flags & REQ_WRITE)) |
1085 | return 0; |
1086 | |
1087 | return atapi_cmd_type(rq->cmd[0]) == ATAPI_MISC; |
1088 | } |
1089 | |
1090 | static int ata_scsi_dev_config(struct scsi_device *sdev, |
1091 | struct ata_device *dev) |
1092 | { |
1093 | struct request_queue *q = sdev->request_queue; |
1094 | |
1095 | if (!ata_id_has_unload(dev->id)) |
1096 | dev->flags |= ATA_DFLAG_NO_UNLOAD; |
1097 | |
1098 | /* configure max sectors */ |
1099 | blk_queue_max_hw_sectors(q, dev->max_sectors); |
1100 | |
1101 | if (dev->class == ATA_DEV_ATAPI) { |
1102 | void *buf; |
1103 | |
1104 | sdev->sector_size = ATA_SECT_SIZE; |
1105 | |
1106 | /* set DMA padding */ |
1107 | blk_queue_update_dma_pad(q, ATA_DMA_PAD_SZ - 1); |
1108 | |
1109 | /* configure draining */ |
1110 | buf = kmalloc(ATAPI_MAX_DRAIN, q->bounce_gfp | GFP_KERNEL); |
1111 | if (!buf) { |
1112 | ata_dev_err(dev, "drain buffer allocation failed\n"); |
1113 | return -ENOMEM; |
1114 | } |
1115 | |
1116 | blk_queue_dma_drain(q, atapi_drain_needed, buf, ATAPI_MAX_DRAIN); |
1117 | } else { |
1118 | sdev->sector_size = ata_id_logical_sector_size(dev->id); |
1119 | sdev->manage_start_stop = 1; |
1120 | } |
1121 | |
1122 | /* |
1123 | * ata_pio_sectors() expects buffer for each sector to not cross |
1124 | * page boundary. Enforce it by requiring buffers to be sector |
1125 | * aligned, which works iff sector_size is not larger than |
1126 | * PAGE_SIZE. ATAPI devices also need the alignment as |
1127 | * IDENTIFY_PACKET is executed as ATA_PROT_PIO. |
1128 | */ |
1129 | if (sdev->sector_size > PAGE_SIZE) |
1130 | ata_dev_warn(dev, |
1131 | "sector_size=%u > PAGE_SIZE, PIO may malfunction\n", |
1132 | sdev->sector_size); |
1133 | |
1134 | blk_queue_update_dma_alignment(q, sdev->sector_size - 1); |
1135 | |
1136 | if (dev->flags & ATA_DFLAG_AN) |
1137 | set_bit(SDEV_EVT_MEDIA_CHANGE, sdev->supported_events); |
1138 | |
1139 | if (dev->flags & ATA_DFLAG_NCQ) { |
1140 | int depth; |
1141 | |
1142 | depth = min(sdev->host->can_queue, ata_id_queue_depth(dev->id)); |
1143 | depth = min(ATA_MAX_QUEUE - 1, depth); |
1144 | scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG, depth); |
1145 | } |
1146 | |
1147 | blk_queue_flush_queueable(q, false); |
1148 | |
1149 | dev->sdev = sdev; |
1150 | return 0; |
1151 | } |
1152 | |
1153 | /** |
1154 | * ata_scsi_slave_config - Set SCSI device attributes |
1155 | * @sdev: SCSI device to examine |
1156 | * |
1157 | * This is called before we actually start reading |
1158 | * and writing to the device, to configure certain |
1159 | * SCSI mid-layer behaviors. |
1160 | * |
1161 | * LOCKING: |
1162 | * Defined by SCSI layer. We don't really care. |
1163 | */ |
1164 | |
1165 | int ata_scsi_slave_config(struct scsi_device *sdev) |
1166 | { |
1167 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1168 | struct ata_device *dev = __ata_scsi_find_dev(ap, sdev); |
1169 | int rc = 0; |
1170 | |
1171 | ata_scsi_sdev_config(sdev); |
1172 | |
1173 | if (dev) |
1174 | rc = ata_scsi_dev_config(sdev, dev); |
1175 | |
1176 | return rc; |
1177 | } |
1178 | |
1179 | /** |
1180 | * ata_scsi_slave_destroy - SCSI device is about to be destroyed |
1181 | * @sdev: SCSI device to be destroyed |
1182 | * |
1183 | * @sdev is about to be destroyed for hot/warm unplugging. If |
1184 | * this unplugging was initiated by libata as indicated by NULL |
1185 | * dev->sdev, this function doesn't have to do anything. |
1186 | * Otherwise, SCSI layer initiated warm-unplug is in progress. |
1187 | * Clear dev->sdev, schedule the device for ATA detach and invoke |
1188 | * EH. |
1189 | * |
1190 | * LOCKING: |
1191 | * Defined by SCSI layer. We don't really care. |
1192 | */ |
1193 | void ata_scsi_slave_destroy(struct scsi_device *sdev) |
1194 | { |
1195 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1196 | struct request_queue *q = sdev->request_queue; |
1197 | unsigned long flags; |
1198 | struct ata_device *dev; |
1199 | |
1200 | if (!ap->ops->error_handler) |
1201 | return; |
1202 | |
1203 | spin_lock_irqsave(ap->lock, flags); |
1204 | dev = __ata_scsi_find_dev(ap, sdev); |
1205 | if (dev && dev->sdev) { |
1206 | /* SCSI device already in CANCEL state, no need to offline it */ |
1207 | dev->sdev = NULL; |
1208 | dev->flags |= ATA_DFLAG_DETACH; |
1209 | ata_port_schedule_eh(ap); |
1210 | } |
1211 | spin_unlock_irqrestore(ap->lock, flags); |
1212 | |
1213 | kfree(q->dma_drain_buffer); |
1214 | q->dma_drain_buffer = NULL; |
1215 | q->dma_drain_size = 0; |
1216 | } |
1217 | |
1218 | /** |
1219 | * __ata_change_queue_depth - helper for ata_scsi_change_queue_depth |
1220 | * @ap: ATA port to which the device change the queue depth |
1221 | * @sdev: SCSI device to configure queue depth for |
1222 | * @queue_depth: new queue depth |
1223 | * @reason: calling context |
1224 | * |
1225 | * libsas and libata have different approaches for associating a sdev to |
1226 | * its ata_port. |
1227 | * |
1228 | */ |
1229 | int __ata_change_queue_depth(struct ata_port *ap, struct scsi_device *sdev, |
1230 | int queue_depth, int reason) |
1231 | { |
1232 | struct ata_device *dev; |
1233 | unsigned long flags; |
1234 | |
1235 | if (reason != SCSI_QDEPTH_DEFAULT) |
1236 | return -EOPNOTSUPP; |
1237 | |
1238 | if (queue_depth < 1 || queue_depth == sdev->queue_depth) |
1239 | return sdev->queue_depth; |
1240 | |
1241 | dev = ata_scsi_find_dev(ap, sdev); |
1242 | if (!dev || !ata_dev_enabled(dev)) |
1243 | return sdev->queue_depth; |
1244 | |
1245 | /* NCQ enabled? */ |
1246 | spin_lock_irqsave(ap->lock, flags); |
1247 | dev->flags &= ~ATA_DFLAG_NCQ_OFF; |
1248 | if (queue_depth == 1 || !ata_ncq_enabled(dev)) { |
1249 | dev->flags |= ATA_DFLAG_NCQ_OFF; |
1250 | queue_depth = 1; |
1251 | } |
1252 | spin_unlock_irqrestore(ap->lock, flags); |
1253 | |
1254 | /* limit and apply queue depth */ |
1255 | queue_depth = min(queue_depth, sdev->host->can_queue); |
1256 | queue_depth = min(queue_depth, ata_id_queue_depth(dev->id)); |
1257 | queue_depth = min(queue_depth, ATA_MAX_QUEUE - 1); |
1258 | |
1259 | if (sdev->queue_depth == queue_depth) |
1260 | return -EINVAL; |
1261 | |
1262 | scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG, queue_depth); |
1263 | return queue_depth; |
1264 | } |
1265 | |
1266 | /** |
1267 | * ata_scsi_change_queue_depth - SCSI callback for queue depth config |
1268 | * @sdev: SCSI device to configure queue depth for |
1269 | * @queue_depth: new queue depth |
1270 | * @reason: calling context |
1271 | * |
1272 | * This is libata standard hostt->change_queue_depth callback. |
1273 | * SCSI will call into this callback when user tries to set queue |
1274 | * depth via sysfs. |
1275 | * |
1276 | * LOCKING: |
1277 | * SCSI layer (we don't care) |
1278 | * |
1279 | * RETURNS: |
1280 | * Newly configured queue depth. |
1281 | */ |
1282 | int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth, |
1283 | int reason) |
1284 | { |
1285 | struct ata_port *ap = ata_shost_to_port(sdev->host); |
1286 | |
1287 | return __ata_change_queue_depth(ap, sdev, queue_depth, reason); |
1288 | } |
1289 | |
1290 | /** |
1291 | * ata_scsi_start_stop_xlat - Translate SCSI START STOP UNIT command |
1292 | * @qc: Storage for translated ATA taskfile |
1293 | * |
1294 | * Sets up an ATA taskfile to issue STANDBY (to stop) or READ VERIFY |
1295 | * (to start). Perhaps these commands should be preceded by |
1296 | * CHECK POWER MODE to see what power mode the device is already in. |
1297 | * [See SAT revision 5 at www.t10.org] |
1298 | * |
1299 | * LOCKING: |
1300 | * spin_lock_irqsave(host lock) |
1301 | * |
1302 | * RETURNS: |
1303 | * Zero on success, non-zero on error. |
1304 | */ |
1305 | static unsigned int ata_scsi_start_stop_xlat(struct ata_queued_cmd *qc) |
1306 | { |
1307 | struct scsi_cmnd *scmd = qc->scsicmd; |
1308 | struct ata_taskfile *tf = &qc->tf; |
1309 | const u8 *cdb = scmd->cmnd; |
1310 | |
1311 | if (scmd->cmd_len < 5) |
1312 | goto invalid_fld; |
1313 | |
1314 | tf->flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; |
1315 | tf->protocol = ATA_PROT_NODATA; |
1316 | if (cdb[1] & 0x1) { |
1317 | ; /* ignore IMMED bit, violates sat-r05 */ |
1318 | } |
1319 | if (cdb[4] & 0x2) |
1320 | goto invalid_fld; /* LOEJ bit set not supported */ |
1321 | if (((cdb[4] >> 4) & 0xf) != 0) |
1322 | goto invalid_fld; /* power conditions not supported */ |
1323 | |
1324 | if (cdb[4] & 0x1) { |
1325 | tf->nsect = 1; /* 1 sector, lba=0 */ |
1326 | |
1327 | if (qc->dev->flags & ATA_DFLAG_LBA) { |
1328 | tf->flags |= ATA_TFLAG_LBA; |
1329 | |
1330 | tf->lbah = 0x0; |
1331 | tf->lbam = 0x0; |
1332 | tf->lbal = 0x0; |
1333 | tf->device |= ATA_LBA; |
1334 | } else { |
1335 | /* CHS */ |
1336 | tf->lbal = 0x1; /* sect */ |
1337 | tf->lbam = 0x0; /* cyl low */ |
1338 | tf->lbah = 0x0; /* cyl high */ |
1339 | } |
1340 | |
1341 | tf->command = ATA_CMD_VERIFY; /* READ VERIFY */ |
1342 | } else { |
1343 | /* Some odd clown BIOSen issue spindown on power off (ACPI S4 |
1344 | * or S5) causing some drives to spin up and down again. |
1345 | */ |
1346 | if ((qc->ap->flags & ATA_FLAG_NO_POWEROFF_SPINDOWN) && |
1347 | system_state == SYSTEM_POWER_OFF) |
1348 | goto skip; |
1349 | |
1350 | if ((qc->ap->flags & ATA_FLAG_NO_HIBERNATE_SPINDOWN) && |
1351 | system_entering_hibernation()) |
1352 | goto skip; |
1353 | |
1354 | /* Issue ATA STANDBY IMMEDIATE command */ |
1355 | tf->command = ATA_CMD_STANDBYNOW1; |
1356 | } |
1357 | |
1358 | /* |
1359 | * Standby and Idle condition timers could be implemented but that |
1360 | * would require libata to implement the Power condition mode page |
1361 | * and allow the user to change it. Changing mode pages requires |
1362 | * MODE SELECT to be implemented. |
1363 | */ |
1364 | |
1365 | return 0; |
1366 | |
1367 | invalid_fld: |
1368 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1369 | /* "Invalid field in cbd" */ |
1370 | return 1; |
1371 | skip: |
1372 | scmd->result = SAM_STAT_GOOD; |
1373 | return 1; |
1374 | } |
1375 | |
1376 | |
1377 | /** |
1378 | * ata_scsi_flush_xlat - Translate SCSI SYNCHRONIZE CACHE command |
1379 | * @qc: Storage for translated ATA taskfile |
1380 | * |
1381 | * Sets up an ATA taskfile to issue FLUSH CACHE or |
1382 | * FLUSH CACHE EXT. |
1383 | * |
1384 | * LOCKING: |
1385 | * spin_lock_irqsave(host lock) |
1386 | * |
1387 | * RETURNS: |
1388 | * Zero on success, non-zero on error. |
1389 | */ |
1390 | static unsigned int ata_scsi_flush_xlat(struct ata_queued_cmd *qc) |
1391 | { |
1392 | struct ata_taskfile *tf = &qc->tf; |
1393 | |
1394 | tf->flags |= ATA_TFLAG_DEVICE; |
1395 | tf->protocol = ATA_PROT_NODATA; |
1396 | |
1397 | if (qc->dev->flags & ATA_DFLAG_FLUSH_EXT) |
1398 | tf->command = ATA_CMD_FLUSH_EXT; |
1399 | else |
1400 | tf->command = ATA_CMD_FLUSH; |
1401 | |
1402 | /* flush is critical for IO integrity, consider it an IO command */ |
1403 | qc->flags |= ATA_QCFLAG_IO; |
1404 | |
1405 | return 0; |
1406 | } |
1407 | |
1408 | /** |
1409 | * scsi_6_lba_len - Get LBA and transfer length |
1410 | * @cdb: SCSI command to translate |
1411 | * |
1412 | * Calculate LBA and transfer length for 6-byte commands. |
1413 | * |
1414 | * RETURNS: |
1415 | * @plba: the LBA |
1416 | * @plen: the transfer length |
1417 | */ |
1418 | static void scsi_6_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1419 | { |
1420 | u64 lba = 0; |
1421 | u32 len; |
1422 | |
1423 | VPRINTK("six-byte command\n"); |
1424 | |
1425 | lba |= ((u64)(cdb[1] & 0x1f)) << 16; |
1426 | lba |= ((u64)cdb[2]) << 8; |
1427 | lba |= ((u64)cdb[3]); |
1428 | |
1429 | len = cdb[4]; |
1430 | |
1431 | *plba = lba; |
1432 | *plen = len; |
1433 | } |
1434 | |
1435 | /** |
1436 | * scsi_10_lba_len - Get LBA and transfer length |
1437 | * @cdb: SCSI command to translate |
1438 | * |
1439 | * Calculate LBA and transfer length for 10-byte commands. |
1440 | * |
1441 | * RETURNS: |
1442 | * @plba: the LBA |
1443 | * @plen: the transfer length |
1444 | */ |
1445 | static void scsi_10_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1446 | { |
1447 | u64 lba = 0; |
1448 | u32 len = 0; |
1449 | |
1450 | VPRINTK("ten-byte command\n"); |
1451 | |
1452 | lba |= ((u64)cdb[2]) << 24; |
1453 | lba |= ((u64)cdb[3]) << 16; |
1454 | lba |= ((u64)cdb[4]) << 8; |
1455 | lba |= ((u64)cdb[5]); |
1456 | |
1457 | len |= ((u32)cdb[7]) << 8; |
1458 | len |= ((u32)cdb[8]); |
1459 | |
1460 | *plba = lba; |
1461 | *plen = len; |
1462 | } |
1463 | |
1464 | /** |
1465 | * scsi_16_lba_len - Get LBA and transfer length |
1466 | * @cdb: SCSI command to translate |
1467 | * |
1468 | * Calculate LBA and transfer length for 16-byte commands. |
1469 | * |
1470 | * RETURNS: |
1471 | * @plba: the LBA |
1472 | * @plen: the transfer length |
1473 | */ |
1474 | static void scsi_16_lba_len(const u8 *cdb, u64 *plba, u32 *plen) |
1475 | { |
1476 | u64 lba = 0; |
1477 | u32 len = 0; |
1478 | |
1479 | VPRINTK("sixteen-byte command\n"); |
1480 | |
1481 | lba |= ((u64)cdb[2]) << 56; |
1482 | lba |= ((u64)cdb[3]) << 48; |
1483 | lba |= ((u64)cdb[4]) << 40; |
1484 | lba |= ((u64)cdb[5]) << 32; |
1485 | lba |= ((u64)cdb[6]) << 24; |
1486 | lba |= ((u64)cdb[7]) << 16; |
1487 | lba |= ((u64)cdb[8]) << 8; |
1488 | lba |= ((u64)cdb[9]); |
1489 | |
1490 | len |= ((u32)cdb[10]) << 24; |
1491 | len |= ((u32)cdb[11]) << 16; |
1492 | len |= ((u32)cdb[12]) << 8; |
1493 | len |= ((u32)cdb[13]); |
1494 | |
1495 | *plba = lba; |
1496 | *plen = len; |
1497 | } |
1498 | |
1499 | /** |
1500 | * ata_scsi_verify_xlat - Translate SCSI VERIFY command into an ATA one |
1501 | * @qc: Storage for translated ATA taskfile |
1502 | * |
1503 | * Converts SCSI VERIFY command to an ATA READ VERIFY command. |
1504 | * |
1505 | * LOCKING: |
1506 | * spin_lock_irqsave(host lock) |
1507 | * |
1508 | * RETURNS: |
1509 | * Zero on success, non-zero on error. |
1510 | */ |
1511 | static unsigned int ata_scsi_verify_xlat(struct ata_queued_cmd *qc) |
1512 | { |
1513 | struct scsi_cmnd *scmd = qc->scsicmd; |
1514 | struct ata_taskfile *tf = &qc->tf; |
1515 | struct ata_device *dev = qc->dev; |
1516 | u64 dev_sectors = qc->dev->n_sectors; |
1517 | const u8 *cdb = scmd->cmnd; |
1518 | u64 block; |
1519 | u32 n_block; |
1520 | |
1521 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
1522 | tf->protocol = ATA_PROT_NODATA; |
1523 | |
1524 | if (cdb[0] == VERIFY) { |
1525 | if (scmd->cmd_len < 10) |
1526 | goto invalid_fld; |
1527 | scsi_10_lba_len(cdb, &block, &n_block); |
1528 | } else if (cdb[0] == VERIFY_16) { |
1529 | if (scmd->cmd_len < 16) |
1530 | goto invalid_fld; |
1531 | scsi_16_lba_len(cdb, &block, &n_block); |
1532 | } else |
1533 | goto invalid_fld; |
1534 | |
1535 | if (!n_block) |
1536 | goto nothing_to_do; |
1537 | if (block >= dev_sectors) |
1538 | goto out_of_range; |
1539 | if ((block + n_block) > dev_sectors) |
1540 | goto out_of_range; |
1541 | |
1542 | if (dev->flags & ATA_DFLAG_LBA) { |
1543 | tf->flags |= ATA_TFLAG_LBA; |
1544 | |
1545 | if (lba_28_ok(block, n_block)) { |
1546 | /* use LBA28 */ |
1547 | tf->command = ATA_CMD_VERIFY; |
1548 | tf->device |= (block >> 24) & 0xf; |
1549 | } else if (lba_48_ok(block, n_block)) { |
1550 | if (!(dev->flags & ATA_DFLAG_LBA48)) |
1551 | goto out_of_range; |
1552 | |
1553 | /* use LBA48 */ |
1554 | tf->flags |= ATA_TFLAG_LBA48; |
1555 | tf->command = ATA_CMD_VERIFY_EXT; |
1556 | |
1557 | tf->hob_nsect = (n_block >> 8) & 0xff; |
1558 | |
1559 | tf->hob_lbah = (block >> 40) & 0xff; |
1560 | tf->hob_lbam = (block >> 32) & 0xff; |
1561 | tf->hob_lbal = (block >> 24) & 0xff; |
1562 | } else |
1563 | /* request too large even for LBA48 */ |
1564 | goto out_of_range; |
1565 | |
1566 | tf->nsect = n_block & 0xff; |
1567 | |
1568 | tf->lbah = (block >> 16) & 0xff; |
1569 | tf->lbam = (block >> 8) & 0xff; |
1570 | tf->lbal = block & 0xff; |
1571 | |
1572 | tf->device |= ATA_LBA; |
1573 | } else { |
1574 | /* CHS */ |
1575 | u32 sect, head, cyl, track; |
1576 | |
1577 | if (!lba_28_ok(block, n_block)) |
1578 | goto out_of_range; |
1579 | |
1580 | /* Convert LBA to CHS */ |
1581 | track = (u32)block / dev->sectors; |
1582 | cyl = track / dev->heads; |
1583 | head = track % dev->heads; |
1584 | sect = (u32)block % dev->sectors + 1; |
1585 | |
1586 | DPRINTK("block %u track %u cyl %u head %u sect %u\n", |
1587 | (u32)block, track, cyl, head, sect); |
1588 | |
1589 | /* Check whether the converted CHS can fit. |
1590 | Cylinder: 0-65535 |
1591 | Head: 0-15 |
1592 | Sector: 1-255*/ |
1593 | if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect)) |
1594 | goto out_of_range; |
1595 | |
1596 | tf->command = ATA_CMD_VERIFY; |
1597 | tf->nsect = n_block & 0xff; /* Sector count 0 means 256 sectors */ |
1598 | tf->lbal = sect; |
1599 | tf->lbam = cyl; |
1600 | tf->lbah = cyl >> 8; |
1601 | tf->device |= head; |
1602 | } |
1603 | |
1604 | return 0; |
1605 | |
1606 | invalid_fld: |
1607 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1608 | /* "Invalid field in cbd" */ |
1609 | return 1; |
1610 | |
1611 | out_of_range: |
1612 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); |
1613 | /* "Logical Block Address out of range" */ |
1614 | return 1; |
1615 | |
1616 | nothing_to_do: |
1617 | scmd->result = SAM_STAT_GOOD; |
1618 | return 1; |
1619 | } |
1620 | |
1621 | /** |
1622 | * ata_scsi_rw_xlat - Translate SCSI r/w command into an ATA one |
1623 | * @qc: Storage for translated ATA taskfile |
1624 | * |
1625 | * Converts any of six SCSI read/write commands into the |
1626 | * ATA counterpart, including starting sector (LBA), |
1627 | * sector count, and taking into account the device's LBA48 |
1628 | * support. |
1629 | * |
1630 | * Commands %READ_6, %READ_10, %READ_16, %WRITE_6, %WRITE_10, and |
1631 | * %WRITE_16 are currently supported. |
1632 | * |
1633 | * LOCKING: |
1634 | * spin_lock_irqsave(host lock) |
1635 | * |
1636 | * RETURNS: |
1637 | * Zero on success, non-zero on error. |
1638 | */ |
1639 | static unsigned int ata_scsi_rw_xlat(struct ata_queued_cmd *qc) |
1640 | { |
1641 | struct scsi_cmnd *scmd = qc->scsicmd; |
1642 | const u8 *cdb = scmd->cmnd; |
1643 | unsigned int tf_flags = 0; |
1644 | u64 block; |
1645 | u32 n_block; |
1646 | int rc; |
1647 | |
1648 | if (cdb[0] == WRITE_10 || cdb[0] == WRITE_6 || cdb[0] == WRITE_16) |
1649 | tf_flags |= ATA_TFLAG_WRITE; |
1650 | |
1651 | /* Calculate the SCSI LBA, transfer length and FUA. */ |
1652 | switch (cdb[0]) { |
1653 | case READ_10: |
1654 | case WRITE_10: |
1655 | if (unlikely(scmd->cmd_len < 10)) |
1656 | goto invalid_fld; |
1657 | scsi_10_lba_len(cdb, &block, &n_block); |
1658 | if (unlikely(cdb[1] & (1 << 3))) |
1659 | tf_flags |= ATA_TFLAG_FUA; |
1660 | break; |
1661 | case READ_6: |
1662 | case WRITE_6: |
1663 | if (unlikely(scmd->cmd_len < 6)) |
1664 | goto invalid_fld; |
1665 | scsi_6_lba_len(cdb, &block, &n_block); |
1666 | |
1667 | /* for 6-byte r/w commands, transfer length 0 |
1668 | * means 256 blocks of data, not 0 block. |
1669 | */ |
1670 | if (!n_block) |
1671 | n_block = 256; |
1672 | break; |
1673 | case READ_16: |
1674 | case WRITE_16: |
1675 | if (unlikely(scmd->cmd_len < 16)) |
1676 | goto invalid_fld; |
1677 | scsi_16_lba_len(cdb, &block, &n_block); |
1678 | if (unlikely(cdb[1] & (1 << 3))) |
1679 | tf_flags |= ATA_TFLAG_FUA; |
1680 | break; |
1681 | default: |
1682 | DPRINTK("no-byte command\n"); |
1683 | goto invalid_fld; |
1684 | } |
1685 | |
1686 | /* Check and compose ATA command */ |
1687 | if (!n_block) |
1688 | /* For 10-byte and 16-byte SCSI R/W commands, transfer |
1689 | * length 0 means transfer 0 block of data. |
1690 | * However, for ATA R/W commands, sector count 0 means |
1691 | * 256 or 65536 sectors, not 0 sectors as in SCSI. |
1692 | * |
1693 | * WARNING: one or two older ATA drives treat 0 as 0... |
1694 | */ |
1695 | goto nothing_to_do; |
1696 | |
1697 | qc->flags |= ATA_QCFLAG_IO; |
1698 | qc->nbytes = n_block * scmd->device->sector_size; |
1699 | |
1700 | rc = ata_build_rw_tf(&qc->tf, qc->dev, block, n_block, tf_flags, |
1701 | qc->tag); |
1702 | if (likely(rc == 0)) |
1703 | return 0; |
1704 | |
1705 | if (rc == -ERANGE) |
1706 | goto out_of_range; |
1707 | /* treat all other errors as -EINVAL, fall through */ |
1708 | invalid_fld: |
1709 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x0); |
1710 | /* "Invalid field in cbd" */ |
1711 | return 1; |
1712 | |
1713 | out_of_range: |
1714 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x21, 0x0); |
1715 | /* "Logical Block Address out of range" */ |
1716 | return 1; |
1717 | |
1718 | nothing_to_do: |
1719 | scmd->result = SAM_STAT_GOOD; |
1720 | return 1; |
1721 | } |
1722 | |
1723 | static void ata_scsi_qc_complete(struct ata_queued_cmd *qc) |
1724 | { |
1725 | struct ata_port *ap = qc->ap; |
1726 | struct scsi_cmnd *cmd = qc->scsicmd; |
1727 | u8 *cdb = cmd->cmnd; |
1728 | int need_sense = (qc->err_mask != 0); |
1729 | |
1730 | /* For ATA pass thru (SAT) commands, generate a sense block if |
1731 | * user mandated it or if there's an error. Note that if we |
1732 | * generate because the user forced us to, a check condition |
1733 | * is generated and the ATA register values are returned |
1734 | * whether the command completed successfully or not. If there |
1735 | * was no error, SK, ASC and ASCQ will all be zero. |
1736 | */ |
1737 | if (((cdb[0] == ATA_16) || (cdb[0] == ATA_12)) && |
1738 | ((cdb[2] & 0x20) || need_sense)) { |
1739 | ata_gen_passthru_sense(qc); |
1740 | } else { |
1741 | if (!need_sense) { |
1742 | cmd->result = SAM_STAT_GOOD; |
1743 | } else { |
1744 | /* TODO: decide which descriptor format to use |
1745 | * for 48b LBA devices and call that here |
1746 | * instead of the fixed desc, which is only |
1747 | * good for smaller LBA (and maybe CHS?) |
1748 | * devices. |
1749 | */ |
1750 | ata_gen_ata_sense(qc); |
1751 | } |
1752 | } |
1753 | |
1754 | if (need_sense && !ap->ops->error_handler) |
1755 | ata_dump_status(ap->print_id, &qc->result_tf); |
1756 | |
1757 | qc->scsidone(cmd); |
1758 | |
1759 | ata_qc_free(qc); |
1760 | } |
1761 | |
1762 | /** |
1763 | * ata_scsi_translate - Translate then issue SCSI command to ATA device |
1764 | * @dev: ATA device to which the command is addressed |
1765 | * @cmd: SCSI command to execute |
1766 | * @xlat_func: Actor which translates @cmd to an ATA taskfile |
1767 | * |
1768 | * Our ->queuecommand() function has decided that the SCSI |
1769 | * command issued can be directly translated into an ATA |
1770 | * command, rather than handled internally. |
1771 | * |
1772 | * This function sets up an ata_queued_cmd structure for the |
1773 | * SCSI command, and sends that ata_queued_cmd to the hardware. |
1774 | * |
1775 | * The xlat_func argument (actor) returns 0 if ready to execute |
1776 | * ATA command, else 1 to finish translation. If 1 is returned |
1777 | * then cmd->result (and possibly cmd->sense_buffer) are assumed |
1778 | * to be set reflecting an error condition or clean (early) |
1779 | * termination. |
1780 | * |
1781 | * LOCKING: |
1782 | * spin_lock_irqsave(host lock) |
1783 | * |
1784 | * RETURNS: |
1785 | * 0 on success, SCSI_ML_QUEUE_DEVICE_BUSY if the command |
1786 | * needs to be deferred. |
1787 | */ |
1788 | static int ata_scsi_translate(struct ata_device *dev, struct scsi_cmnd *cmd, |
1789 | ata_xlat_func_t xlat_func) |
1790 | { |
1791 | struct ata_port *ap = dev->link->ap; |
1792 | struct ata_queued_cmd *qc; |
1793 | int rc; |
1794 | |
1795 | VPRINTK("ENTER\n"); |
1796 | |
1797 | qc = ata_scsi_qc_new(dev, cmd); |
1798 | if (!qc) |
1799 | goto err_mem; |
1800 | |
1801 | /* data is present; dma-map it */ |
1802 | if (cmd->sc_data_direction == DMA_FROM_DEVICE || |
1803 | cmd->sc_data_direction == DMA_TO_DEVICE) { |
1804 | if (unlikely(scsi_bufflen(cmd) < 1)) { |
1805 | ata_dev_warn(dev, "WARNING: zero len r/w req\n"); |
1806 | goto err_did; |
1807 | } |
1808 | |
1809 | ata_sg_init(qc, scsi_sglist(cmd), scsi_sg_count(cmd)); |
1810 | |
1811 | qc->dma_dir = cmd->sc_data_direction; |
1812 | } |
1813 | |
1814 | qc->complete_fn = ata_scsi_qc_complete; |
1815 | |
1816 | if (xlat_func(qc)) |
1817 | goto early_finish; |
1818 | |
1819 | if (ap->ops->qc_defer) { |
1820 | if ((rc = ap->ops->qc_defer(qc))) |
1821 | goto defer; |
1822 | } |
1823 | |
1824 | /* select device, send command to hardware */ |
1825 | ata_qc_issue(qc); |
1826 | |
1827 | VPRINTK("EXIT\n"); |
1828 | return 0; |
1829 | |
1830 | early_finish: |
1831 | ata_qc_free(qc); |
1832 | cmd->scsi_done(cmd); |
1833 | DPRINTK("EXIT - early finish (good or error)\n"); |
1834 | return 0; |
1835 | |
1836 | err_did: |
1837 | ata_qc_free(qc); |
1838 | cmd->result = (DID_ERROR << 16); |
1839 | cmd->scsi_done(cmd); |
1840 | err_mem: |
1841 | DPRINTK("EXIT - internal\n"); |
1842 | return 0; |
1843 | |
1844 | defer: |
1845 | ata_qc_free(qc); |
1846 | DPRINTK("EXIT - defer\n"); |
1847 | if (rc == ATA_DEFER_LINK) |
1848 | return SCSI_MLQUEUE_DEVICE_BUSY; |
1849 | else |
1850 | return SCSI_MLQUEUE_HOST_BUSY; |
1851 | } |
1852 | |
1853 | /** |
1854 | * ata_scsi_rbuf_get - Map response buffer. |
1855 | * @cmd: SCSI command containing buffer to be mapped. |
1856 | * @flags: unsigned long variable to store irq enable status |
1857 | * @copy_in: copy in from user buffer |
1858 | * |
1859 | * Prepare buffer for simulated SCSI commands. |
1860 | * |
1861 | * LOCKING: |
1862 | * spin_lock_irqsave(ata_scsi_rbuf_lock) on success |
1863 | * |
1864 | * RETURNS: |
1865 | * Pointer to response buffer. |
1866 | */ |
1867 | static void *ata_scsi_rbuf_get(struct scsi_cmnd *cmd, bool copy_in, |
1868 | unsigned long *flags) |
1869 | { |
1870 | spin_lock_irqsave(&ata_scsi_rbuf_lock, *flags); |
1871 | |
1872 | memset(ata_scsi_rbuf, 0, ATA_SCSI_RBUF_SIZE); |
1873 | if (copy_in) |
1874 | sg_copy_to_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), |
1875 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); |
1876 | return ata_scsi_rbuf; |
1877 | } |
1878 | |
1879 | /** |
1880 | * ata_scsi_rbuf_put - Unmap response buffer. |
1881 | * @cmd: SCSI command containing buffer to be unmapped. |
1882 | * @copy_out: copy out result |
1883 | * @flags: @flags passed to ata_scsi_rbuf_get() |
1884 | * |
1885 | * Returns rbuf buffer. The result is copied to @cmd's buffer if |
1886 | * @copy_back is true. |
1887 | * |
1888 | * LOCKING: |
1889 | * Unlocks ata_scsi_rbuf_lock. |
1890 | */ |
1891 | static inline void ata_scsi_rbuf_put(struct scsi_cmnd *cmd, bool copy_out, |
1892 | unsigned long *flags) |
1893 | { |
1894 | if (copy_out) |
1895 | sg_copy_from_buffer(scsi_sglist(cmd), scsi_sg_count(cmd), |
1896 | ata_scsi_rbuf, ATA_SCSI_RBUF_SIZE); |
1897 | spin_unlock_irqrestore(&ata_scsi_rbuf_lock, *flags); |
1898 | } |
1899 | |
1900 | /** |
1901 | * ata_scsi_rbuf_fill - wrapper for SCSI command simulators |
1902 | * @args: device IDENTIFY data / SCSI command of interest. |
1903 | * @actor: Callback hook for desired SCSI command simulator |
1904 | * |
1905 | * Takes care of the hard work of simulating a SCSI command... |
1906 | * Mapping the response buffer, calling the command's handler, |
1907 | * and handling the handler's return value. This return value |
1908 | * indicates whether the handler wishes the SCSI command to be |
1909 | * completed successfully (0), or not (in which case cmd->result |
1910 | * and sense buffer are assumed to be set). |
1911 | * |
1912 | * LOCKING: |
1913 | * spin_lock_irqsave(host lock) |
1914 | */ |
1915 | static void ata_scsi_rbuf_fill(struct ata_scsi_args *args, |
1916 | unsigned int (*actor)(struct ata_scsi_args *args, u8 *rbuf)) |
1917 | { |
1918 | u8 *rbuf; |
1919 | unsigned int rc; |
1920 | struct scsi_cmnd *cmd = args->cmd; |
1921 | unsigned long flags; |
1922 | |
1923 | rbuf = ata_scsi_rbuf_get(cmd, false, &flags); |
1924 | rc = actor(args, rbuf); |
1925 | ata_scsi_rbuf_put(cmd, rc == 0, &flags); |
1926 | |
1927 | if (rc == 0) |
1928 | cmd->result = SAM_STAT_GOOD; |
1929 | args->done(cmd); |
1930 | } |
1931 | |
1932 | /** |
1933 | * ata_scsiop_inq_std - Simulate INQUIRY command |
1934 | * @args: device IDENTIFY data / SCSI command of interest. |
1935 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
1936 | * |
1937 | * Returns standard device identification data associated |
1938 | * with non-VPD INQUIRY command output. |
1939 | * |
1940 | * LOCKING: |
1941 | * spin_lock_irqsave(host lock) |
1942 | */ |
1943 | static unsigned int ata_scsiop_inq_std(struct ata_scsi_args *args, u8 *rbuf) |
1944 | { |
1945 | const u8 versions[] = { |
1946 | 0x60, /* SAM-3 (no version claimed) */ |
1947 | |
1948 | 0x03, |
1949 | 0x20, /* SBC-2 (no version claimed) */ |
1950 | |
1951 | 0x02, |
1952 | 0x60 /* SPC-3 (no version claimed) */ |
1953 | }; |
1954 | u8 hdr[] = { |
1955 | TYPE_DISK, |
1956 | 0, |
1957 | 0x5, /* claim SPC-3 version compatibility */ |
1958 | 2, |
1959 | 95 - 4 |
1960 | }; |
1961 | |
1962 | VPRINTK("ENTER\n"); |
1963 | |
1964 | /* set scsi removeable (RMB) bit per ata bit */ |
1965 | if (ata_id_removeable(args->id)) |
1966 | hdr[1] |= (1 << 7); |
1967 | |
1968 | memcpy(rbuf, hdr, sizeof(hdr)); |
1969 | memcpy(&rbuf[8], "ATA ", 8); |
1970 | ata_id_string(args->id, &rbuf[16], ATA_ID_PROD, 16); |
1971 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV, 4); |
1972 | |
1973 | if (rbuf[32] == 0 || rbuf[32] == ' ') |
1974 | memcpy(&rbuf[32], "n/a ", 4); |
1975 | |
1976 | memcpy(rbuf + 59, versions, sizeof(versions)); |
1977 | |
1978 | return 0; |
1979 | } |
1980 | |
1981 | /** |
1982 | * ata_scsiop_inq_00 - Simulate INQUIRY VPD page 0, list of pages |
1983 | * @args: device IDENTIFY data / SCSI command of interest. |
1984 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
1985 | * |
1986 | * Returns list of inquiry VPD pages available. |
1987 | * |
1988 | * LOCKING: |
1989 | * spin_lock_irqsave(host lock) |
1990 | */ |
1991 | static unsigned int ata_scsiop_inq_00(struct ata_scsi_args *args, u8 *rbuf) |
1992 | { |
1993 | const u8 pages[] = { |
1994 | 0x00, /* page 0x00, this page */ |
1995 | 0x80, /* page 0x80, unit serial no page */ |
1996 | 0x83, /* page 0x83, device ident page */ |
1997 | 0x89, /* page 0x89, ata info page */ |
1998 | 0xb0, /* page 0xb0, block limits page */ |
1999 | 0xb1, /* page 0xb1, block device characteristics page */ |
2000 | 0xb2, /* page 0xb2, thin provisioning page */ |
2001 | }; |
2002 | |
2003 | rbuf[3] = sizeof(pages); /* number of supported VPD pages */ |
2004 | memcpy(rbuf + 4, pages, sizeof(pages)); |
2005 | return 0; |
2006 | } |
2007 | |
2008 | /** |
2009 | * ata_scsiop_inq_80 - Simulate INQUIRY VPD page 80, device serial number |
2010 | * @args: device IDENTIFY data / SCSI command of interest. |
2011 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2012 | * |
2013 | * Returns ATA device serial number. |
2014 | * |
2015 | * LOCKING: |
2016 | * spin_lock_irqsave(host lock) |
2017 | */ |
2018 | static unsigned int ata_scsiop_inq_80(struct ata_scsi_args *args, u8 *rbuf) |
2019 | { |
2020 | const u8 hdr[] = { |
2021 | 0, |
2022 | 0x80, /* this page code */ |
2023 | 0, |
2024 | ATA_ID_SERNO_LEN, /* page len */ |
2025 | }; |
2026 | |
2027 | memcpy(rbuf, hdr, sizeof(hdr)); |
2028 | ata_id_string(args->id, (unsigned char *) &rbuf[4], |
2029 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
2030 | return 0; |
2031 | } |
2032 | |
2033 | /** |
2034 | * ata_scsiop_inq_83 - Simulate INQUIRY VPD page 83, device identity |
2035 | * @args: device IDENTIFY data / SCSI command of interest. |
2036 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2037 | * |
2038 | * Yields two logical unit device identification designators: |
2039 | * - vendor specific ASCII containing the ATA serial number |
2040 | * - SAT defined "t10 vendor id based" containing ASCII vendor |
2041 | * name ("ATA "), model and serial numbers. |
2042 | * |
2043 | * LOCKING: |
2044 | * spin_lock_irqsave(host lock) |
2045 | */ |
2046 | static unsigned int ata_scsiop_inq_83(struct ata_scsi_args *args, u8 *rbuf) |
2047 | { |
2048 | const int sat_model_serial_desc_len = 68; |
2049 | int num; |
2050 | |
2051 | rbuf[1] = 0x83; /* this page code */ |
2052 | num = 4; |
2053 | |
2054 | /* piv=0, assoc=lu, code_set=ACSII, designator=vendor */ |
2055 | rbuf[num + 0] = 2; |
2056 | rbuf[num + 3] = ATA_ID_SERNO_LEN; |
2057 | num += 4; |
2058 | ata_id_string(args->id, (unsigned char *) rbuf + num, |
2059 | ATA_ID_SERNO, ATA_ID_SERNO_LEN); |
2060 | num += ATA_ID_SERNO_LEN; |
2061 | |
2062 | /* SAT defined lu model and serial numbers descriptor */ |
2063 | /* piv=0, assoc=lu, code_set=ACSII, designator=t10 vendor id */ |
2064 | rbuf[num + 0] = 2; |
2065 | rbuf[num + 1] = 1; |
2066 | rbuf[num + 3] = sat_model_serial_desc_len; |
2067 | num += 4; |
2068 | memcpy(rbuf + num, "ATA ", 8); |
2069 | num += 8; |
2070 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_PROD, |
2071 | ATA_ID_PROD_LEN); |
2072 | num += ATA_ID_PROD_LEN; |
2073 | ata_id_string(args->id, (unsigned char *) rbuf + num, ATA_ID_SERNO, |
2074 | ATA_ID_SERNO_LEN); |
2075 | num += ATA_ID_SERNO_LEN; |
2076 | |
2077 | if (ata_id_has_wwn(args->id)) { |
2078 | /* SAT defined lu world wide name */ |
2079 | /* piv=0, assoc=lu, code_set=binary, designator=NAA */ |
2080 | rbuf[num + 0] = 1; |
2081 | rbuf[num + 1] = 3; |
2082 | rbuf[num + 3] = ATA_ID_WWN_LEN; |
2083 | num += 4; |
2084 | ata_id_string(args->id, (unsigned char *) rbuf + num, |
2085 | ATA_ID_WWN, ATA_ID_WWN_LEN); |
2086 | num += ATA_ID_WWN_LEN; |
2087 | } |
2088 | rbuf[3] = num - 4; /* page len (assume less than 256 bytes) */ |
2089 | return 0; |
2090 | } |
2091 | |
2092 | /** |
2093 | * ata_scsiop_inq_89 - Simulate INQUIRY VPD page 89, ATA info |
2094 | * @args: device IDENTIFY data / SCSI command of interest. |
2095 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2096 | * |
2097 | * Yields SAT-specified ATA VPD page. |
2098 | * |
2099 | * LOCKING: |
2100 | * spin_lock_irqsave(host lock) |
2101 | */ |
2102 | static unsigned int ata_scsiop_inq_89(struct ata_scsi_args *args, u8 *rbuf) |
2103 | { |
2104 | struct ata_taskfile tf; |
2105 | |
2106 | memset(&tf, 0, sizeof(tf)); |
2107 | |
2108 | rbuf[1] = 0x89; /* our page code */ |
2109 | rbuf[2] = (0x238 >> 8); /* page size fixed at 238h */ |
2110 | rbuf[3] = (0x238 & 0xff); |
2111 | |
2112 | memcpy(&rbuf[8], "linux ", 8); |
2113 | memcpy(&rbuf[16], "libata ", 16); |
2114 | memcpy(&rbuf[32], DRV_VERSION, 4); |
2115 | ata_id_string(args->id, &rbuf[32], ATA_ID_FW_REV, 4); |
2116 | |
2117 | /* we don't store the ATA device signature, so we fake it */ |
2118 | |
2119 | tf.command = ATA_DRDY; /* really, this is Status reg */ |
2120 | tf.lbal = 0x1; |
2121 | tf.nsect = 0x1; |
2122 | |
2123 | ata_tf_to_fis(&tf, 0, 1, &rbuf[36]); /* TODO: PMP? */ |
2124 | rbuf[36] = 0x34; /* force D2H Reg FIS (34h) */ |
2125 | |
2126 | rbuf[56] = ATA_CMD_ID_ATA; |
2127 | |
2128 | memcpy(&rbuf[60], &args->id[0], 512); |
2129 | return 0; |
2130 | } |
2131 | |
2132 | static unsigned int ata_scsiop_inq_b0(struct ata_scsi_args *args, u8 *rbuf) |
2133 | { |
2134 | u16 min_io_sectors; |
2135 | |
2136 | rbuf[1] = 0xb0; |
2137 | rbuf[3] = 0x3c; /* required VPD size with unmap support */ |
2138 | |
2139 | /* |
2140 | * Optimal transfer length granularity. |
2141 | * |
2142 | * This is always one physical block, but for disks with a smaller |
2143 | * logical than physical sector size we need to figure out what the |
2144 | * latter is. |
2145 | */ |
2146 | min_io_sectors = 1 << ata_id_log2_per_physical_sector(args->id); |
2147 | put_unaligned_be16(min_io_sectors, &rbuf[6]); |
2148 | |
2149 | /* |
2150 | * Optimal unmap granularity. |
2151 | * |
2152 | * The ATA spec doesn't even know about a granularity or alignment |
2153 | * for the TRIM command. We can leave away most of the unmap related |
2154 | * VPD page entries, but we have specifify a granularity to signal |
2155 | * that we support some form of unmap - in thise case via WRITE SAME |
2156 | * with the unmap bit set. |
2157 | */ |
2158 | if (ata_id_has_trim(args->id)) { |
2159 | put_unaligned_be64(65535 * 512 / 8, &rbuf[36]); |
2160 | put_unaligned_be32(1, &rbuf[28]); |
2161 | } |
2162 | |
2163 | return 0; |
2164 | } |
2165 | |
2166 | static unsigned int ata_scsiop_inq_b1(struct ata_scsi_args *args, u8 *rbuf) |
2167 | { |
2168 | int form_factor = ata_id_form_factor(args->id); |
2169 | int media_rotation_rate = ata_id_rotation_rate(args->id); |
2170 | |
2171 | rbuf[1] = 0xb1; |
2172 | rbuf[3] = 0x3c; |
2173 | rbuf[4] = media_rotation_rate >> 8; |
2174 | rbuf[5] = media_rotation_rate; |
2175 | rbuf[7] = form_factor; |
2176 | |
2177 | return 0; |
2178 | } |
2179 | |
2180 | static unsigned int ata_scsiop_inq_b2(struct ata_scsi_args *args, u8 *rbuf) |
2181 | { |
2182 | /* SCSI Thin Provisioning VPD page: SBC-3 rev 22 or later */ |
2183 | rbuf[1] = 0xb2; |
2184 | rbuf[3] = 0x4; |
2185 | rbuf[5] = 1 << 6; /* TPWS */ |
2186 | |
2187 | return 0; |
2188 | } |
2189 | |
2190 | /** |
2191 | * ata_scsiop_noop - Command handler that simply returns success. |
2192 | * @args: device IDENTIFY data / SCSI command of interest. |
2193 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2194 | * |
2195 | * No operation. Simply returns success to caller, to indicate |
2196 | * that the caller should successfully complete this SCSI command. |
2197 | * |
2198 | * LOCKING: |
2199 | * spin_lock_irqsave(host lock) |
2200 | */ |
2201 | static unsigned int ata_scsiop_noop(struct ata_scsi_args *args, u8 *rbuf) |
2202 | { |
2203 | VPRINTK("ENTER\n"); |
2204 | return 0; |
2205 | } |
2206 | |
2207 | /** |
2208 | * ata_msense_caching - Simulate MODE SENSE caching info page |
2209 | * @id: device IDENTIFY data |
2210 | * @buf: output buffer |
2211 | * |
2212 | * Generate a caching info page, which conditionally indicates |
2213 | * write caching to the SCSI layer, depending on device |
2214 | * capabilities. |
2215 | * |
2216 | * LOCKING: |
2217 | * None. |
2218 | */ |
2219 | static unsigned int ata_msense_caching(u16 *id, u8 *buf) |
2220 | { |
2221 | memcpy(buf, def_cache_mpage, sizeof(def_cache_mpage)); |
2222 | if (ata_id_wcache_enabled(id)) |
2223 | buf[2] |= (1 << 2); /* write cache enable */ |
2224 | if (!ata_id_rahead_enabled(id)) |
2225 | buf[12] |= (1 << 5); /* disable read ahead */ |
2226 | return sizeof(def_cache_mpage); |
2227 | } |
2228 | |
2229 | /** |
2230 | * ata_msense_ctl_mode - Simulate MODE SENSE control mode page |
2231 | * @buf: output buffer |
2232 | * |
2233 | * Generate a generic MODE SENSE control mode page. |
2234 | * |
2235 | * LOCKING: |
2236 | * None. |
2237 | */ |
2238 | static unsigned int ata_msense_ctl_mode(u8 *buf) |
2239 | { |
2240 | memcpy(buf, def_control_mpage, sizeof(def_control_mpage)); |
2241 | return sizeof(def_control_mpage); |
2242 | } |
2243 | |
2244 | /** |
2245 | * ata_msense_rw_recovery - Simulate MODE SENSE r/w error recovery page |
2246 | * @buf: output buffer |
2247 | * |
2248 | * Generate a generic MODE SENSE r/w error recovery page. |
2249 | * |
2250 | * LOCKING: |
2251 | * None. |
2252 | */ |
2253 | static unsigned int ata_msense_rw_recovery(u8 *buf) |
2254 | { |
2255 | memcpy(buf, def_rw_recovery_mpage, sizeof(def_rw_recovery_mpage)); |
2256 | return sizeof(def_rw_recovery_mpage); |
2257 | } |
2258 | |
2259 | /* |
2260 | * We can turn this into a real blacklist if it's needed, for now just |
2261 | * blacklist any Maxtor BANC1G10 revision firmware |
2262 | */ |
2263 | static int ata_dev_supports_fua(u16 *id) |
2264 | { |
2265 | unsigned char model[ATA_ID_PROD_LEN + 1], fw[ATA_ID_FW_REV_LEN + 1]; |
2266 | |
2267 | if (!libata_fua) |
2268 | return 0; |
2269 | if (!ata_id_has_fua(id)) |
2270 | return 0; |
2271 | |
2272 | ata_id_c_string(id, model, ATA_ID_PROD, sizeof(model)); |
2273 | ata_id_c_string(id, fw, ATA_ID_FW_REV, sizeof(fw)); |
2274 | |
2275 | if (strcmp(model, "Maxtor")) |
2276 | return 1; |
2277 | if (strcmp(fw, "BANC1G10")) |
2278 | return 1; |
2279 | |
2280 | return 0; /* blacklisted */ |
2281 | } |
2282 | |
2283 | /** |
2284 | * ata_scsiop_mode_sense - Simulate MODE SENSE 6, 10 commands |
2285 | * @args: device IDENTIFY data / SCSI command of interest. |
2286 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2287 | * |
2288 | * Simulate MODE SENSE commands. Assume this is invoked for direct |
2289 | * access devices (e.g. disks) only. There should be no block |
2290 | * descriptor for other device types. |
2291 | * |
2292 | * LOCKING: |
2293 | * spin_lock_irqsave(host lock) |
2294 | */ |
2295 | static unsigned int ata_scsiop_mode_sense(struct ata_scsi_args *args, u8 *rbuf) |
2296 | { |
2297 | struct ata_device *dev = args->dev; |
2298 | u8 *scsicmd = args->cmd->cmnd, *p = rbuf; |
2299 | const u8 sat_blk_desc[] = { |
2300 | 0, 0, 0, 0, /* number of blocks: sat unspecified */ |
2301 | 0, |
2302 | 0, 0x2, 0x0 /* block length: 512 bytes */ |
2303 | }; |
2304 | u8 pg, spg; |
2305 | unsigned int ebd, page_control, six_byte; |
2306 | u8 dpofua; |
2307 | |
2308 | VPRINTK("ENTER\n"); |
2309 | |
2310 | six_byte = (scsicmd[0] == MODE_SENSE); |
2311 | ebd = !(scsicmd[1] & 0x8); /* dbd bit inverted == edb */ |
2312 | /* |
2313 | * LLBA bit in msense(10) ignored (compliant) |
2314 | */ |
2315 | |
2316 | page_control = scsicmd[2] >> 6; |
2317 | switch (page_control) { |
2318 | case 0: /* current */ |
2319 | break; /* supported */ |
2320 | case 3: /* saved */ |
2321 | goto saving_not_supp; |
2322 | case 1: /* changeable */ |
2323 | case 2: /* defaults */ |
2324 | default: |
2325 | goto invalid_fld; |
2326 | } |
2327 | |
2328 | if (six_byte) |
2329 | p += 4 + (ebd ? 8 : 0); |
2330 | else |
2331 | p += 8 + (ebd ? 8 : 0); |
2332 | |
2333 | pg = scsicmd[2] & 0x3f; |
2334 | spg = scsicmd[3]; |
2335 | /* |
2336 | * No mode subpages supported (yet) but asking for _all_ |
2337 | * subpages may be valid |
2338 | */ |
2339 | if (spg && (spg != ALL_SUB_MPAGES)) |
2340 | goto invalid_fld; |
2341 | |
2342 | switch(pg) { |
2343 | case RW_RECOVERY_MPAGE: |
2344 | p += ata_msense_rw_recovery(p); |
2345 | break; |
2346 | |
2347 | case CACHE_MPAGE: |
2348 | p += ata_msense_caching(args->id, p); |
2349 | break; |
2350 | |
2351 | case CONTROL_MPAGE: |
2352 | p += ata_msense_ctl_mode(p); |
2353 | break; |
2354 | |
2355 | case ALL_MPAGES: |
2356 | p += ata_msense_rw_recovery(p); |
2357 | p += ata_msense_caching(args->id, p); |
2358 | p += ata_msense_ctl_mode(p); |
2359 | break; |
2360 | |
2361 | default: /* invalid page code */ |
2362 | goto invalid_fld; |
2363 | } |
2364 | |
2365 | dpofua = 0; |
2366 | if (ata_dev_supports_fua(args->id) && (dev->flags & ATA_DFLAG_LBA48) && |
2367 | (!(dev->flags & ATA_DFLAG_PIO) || dev->multi_count)) |
2368 | dpofua = 1 << 4; |
2369 | |
2370 | if (six_byte) { |
2371 | rbuf[0] = p - rbuf - 1; |
2372 | rbuf[2] |= dpofua; |
2373 | if (ebd) { |
2374 | rbuf[3] = sizeof(sat_blk_desc); |
2375 | memcpy(rbuf + 4, sat_blk_desc, sizeof(sat_blk_desc)); |
2376 | } |
2377 | } else { |
2378 | unsigned int output_len = p - rbuf - 2; |
2379 | |
2380 | rbuf[0] = output_len >> 8; |
2381 | rbuf[1] = output_len; |
2382 | rbuf[3] |= dpofua; |
2383 | if (ebd) { |
2384 | rbuf[7] = sizeof(sat_blk_desc); |
2385 | memcpy(rbuf + 8, sat_blk_desc, sizeof(sat_blk_desc)); |
2386 | } |
2387 | } |
2388 | return 0; |
2389 | |
2390 | invalid_fld: |
2391 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x24, 0x0); |
2392 | /* "Invalid field in cbd" */ |
2393 | return 1; |
2394 | |
2395 | saving_not_supp: |
2396 | ata_scsi_set_sense(args->cmd, ILLEGAL_REQUEST, 0x39, 0x0); |
2397 | /* "Saving parameters not supported" */ |
2398 | return 1; |
2399 | } |
2400 | |
2401 | /** |
2402 | * ata_scsiop_read_cap - Simulate READ CAPACITY[ 16] commands |
2403 | * @args: device IDENTIFY data / SCSI command of interest. |
2404 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2405 | * |
2406 | * Simulate READ CAPACITY commands. |
2407 | * |
2408 | * LOCKING: |
2409 | * None. |
2410 | */ |
2411 | static unsigned int ata_scsiop_read_cap(struct ata_scsi_args *args, u8 *rbuf) |
2412 | { |
2413 | struct ata_device *dev = args->dev; |
2414 | u64 last_lba = dev->n_sectors - 1; /* LBA of the last block */ |
2415 | u32 sector_size; /* physical sector size in bytes */ |
2416 | u8 log2_per_phys; |
2417 | u16 lowest_aligned; |
2418 | |
2419 | sector_size = ata_id_logical_sector_size(dev->id); |
2420 | log2_per_phys = ata_id_log2_per_physical_sector(dev->id); |
2421 | lowest_aligned = ata_id_logical_sector_offset(dev->id, log2_per_phys); |
2422 | |
2423 | VPRINTK("ENTER\n"); |
2424 | |
2425 | if (args->cmd->cmnd[0] == READ_CAPACITY) { |
2426 | if (last_lba >= 0xffffffffULL) |
2427 | last_lba = 0xffffffff; |
2428 | |
2429 | /* sector count, 32-bit */ |
2430 | rbuf[0] = last_lba >> (8 * 3); |
2431 | rbuf[1] = last_lba >> (8 * 2); |
2432 | rbuf[2] = last_lba >> (8 * 1); |
2433 | rbuf[3] = last_lba; |
2434 | |
2435 | /* sector size */ |
2436 | rbuf[4] = sector_size >> (8 * 3); |
2437 | rbuf[5] = sector_size >> (8 * 2); |
2438 | rbuf[6] = sector_size >> (8 * 1); |
2439 | rbuf[7] = sector_size; |
2440 | } else { |
2441 | /* sector count, 64-bit */ |
2442 | rbuf[0] = last_lba >> (8 * 7); |
2443 | rbuf[1] = last_lba >> (8 * 6); |
2444 | rbuf[2] = last_lba >> (8 * 5); |
2445 | rbuf[3] = last_lba >> (8 * 4); |
2446 | rbuf[4] = last_lba >> (8 * 3); |
2447 | rbuf[5] = last_lba >> (8 * 2); |
2448 | rbuf[6] = last_lba >> (8 * 1); |
2449 | rbuf[7] = last_lba; |
2450 | |
2451 | /* sector size */ |
2452 | rbuf[ 8] = sector_size >> (8 * 3); |
2453 | rbuf[ 9] = sector_size >> (8 * 2); |
2454 | rbuf[10] = sector_size >> (8 * 1); |
2455 | rbuf[11] = sector_size; |
2456 | |
2457 | rbuf[12] = 0; |
2458 | rbuf[13] = log2_per_phys; |
2459 | rbuf[14] = (lowest_aligned >> 8) & 0x3f; |
2460 | rbuf[15] = lowest_aligned; |
2461 | |
2462 | if (ata_id_has_trim(args->id)) { |
2463 | rbuf[14] |= 0x80; /* TPE */ |
2464 | |
2465 | if (ata_id_has_zero_after_trim(args->id)) |
2466 | rbuf[14] |= 0x40; /* TPRZ */ |
2467 | } |
2468 | } |
2469 | |
2470 | return 0; |
2471 | } |
2472 | |
2473 | /** |
2474 | * ata_scsiop_report_luns - Simulate REPORT LUNS command |
2475 | * @args: device IDENTIFY data / SCSI command of interest. |
2476 | * @rbuf: Response buffer, to which simulated SCSI cmd output is sent. |
2477 | * |
2478 | * Simulate REPORT LUNS command. |
2479 | * |
2480 | * LOCKING: |
2481 | * spin_lock_irqsave(host lock) |
2482 | */ |
2483 | static unsigned int ata_scsiop_report_luns(struct ata_scsi_args *args, u8 *rbuf) |
2484 | { |
2485 | VPRINTK("ENTER\n"); |
2486 | rbuf[3] = 8; /* just one lun, LUN 0, size 8 bytes */ |
2487 | |
2488 | return 0; |
2489 | } |
2490 | |
2491 | static void atapi_sense_complete(struct ata_queued_cmd *qc) |
2492 | { |
2493 | if (qc->err_mask && ((qc->err_mask & AC_ERR_DEV) == 0)) { |
2494 | /* FIXME: not quite right; we don't want the |
2495 | * translation of taskfile registers into |
2496 | * a sense descriptors, since that's only |
2497 | * correct for ATA, not ATAPI |
2498 | */ |
2499 | ata_gen_passthru_sense(qc); |
2500 | } |
2501 | |
2502 | qc->scsidone(qc->scsicmd); |
2503 | ata_qc_free(qc); |
2504 | } |
2505 | |
2506 | /* is it pointless to prefer PIO for "safety reasons"? */ |
2507 | static inline int ata_pio_use_silly(struct ata_port *ap) |
2508 | { |
2509 | return (ap->flags & ATA_FLAG_PIO_DMA); |
2510 | } |
2511 | |
2512 | static void atapi_request_sense(struct ata_queued_cmd *qc) |
2513 | { |
2514 | struct ata_port *ap = qc->ap; |
2515 | struct scsi_cmnd *cmd = qc->scsicmd; |
2516 | |
2517 | DPRINTK("ATAPI request sense\n"); |
2518 | |
2519 | /* FIXME: is this needed? */ |
2520 | memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); |
2521 | |
2522 | #ifdef CONFIG_ATA_SFF |
2523 | if (ap->ops->sff_tf_read) |
2524 | ap->ops->sff_tf_read(ap, &qc->tf); |
2525 | #endif |
2526 | |
2527 | /* fill these in, for the case where they are -not- overwritten */ |
2528 | cmd->sense_buffer[0] = 0x70; |
2529 | cmd->sense_buffer[2] = qc->tf.feature >> 4; |
2530 | |
2531 | ata_qc_reinit(qc); |
2532 | |
2533 | /* setup sg table and init transfer direction */ |
2534 | sg_init_one(&qc->sgent, cmd->sense_buffer, SCSI_SENSE_BUFFERSIZE); |
2535 | ata_sg_init(qc, &qc->sgent, 1); |
2536 | qc->dma_dir = DMA_FROM_DEVICE; |
2537 | |
2538 | memset(&qc->cdb, 0, qc->dev->cdb_len); |
2539 | qc->cdb[0] = REQUEST_SENSE; |
2540 | qc->cdb[4] = SCSI_SENSE_BUFFERSIZE; |
2541 | |
2542 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
2543 | qc->tf.command = ATA_CMD_PACKET; |
2544 | |
2545 | if (ata_pio_use_silly(ap)) { |
2546 | qc->tf.protocol = ATAPI_PROT_DMA; |
2547 | qc->tf.feature |= ATAPI_PKT_DMA; |
2548 | } else { |
2549 | qc->tf.protocol = ATAPI_PROT_PIO; |
2550 | qc->tf.lbam = SCSI_SENSE_BUFFERSIZE; |
2551 | qc->tf.lbah = 0; |
2552 | } |
2553 | qc->nbytes = SCSI_SENSE_BUFFERSIZE; |
2554 | |
2555 | qc->complete_fn = atapi_sense_complete; |
2556 | |
2557 | ata_qc_issue(qc); |
2558 | |
2559 | DPRINTK("EXIT\n"); |
2560 | } |
2561 | |
2562 | static void atapi_qc_complete(struct ata_queued_cmd *qc) |
2563 | { |
2564 | struct scsi_cmnd *cmd = qc->scsicmd; |
2565 | unsigned int err_mask = qc->err_mask; |
2566 | |
2567 | VPRINTK("ENTER, err_mask 0x%X\n", err_mask); |
2568 | |
2569 | /* handle completion from new EH */ |
2570 | if (unlikely(qc->ap->ops->error_handler && |
2571 | (err_mask || qc->flags & ATA_QCFLAG_SENSE_VALID))) { |
2572 | |
2573 | if (!(qc->flags & ATA_QCFLAG_SENSE_VALID)) { |
2574 | /* FIXME: not quite right; we don't want the |
2575 | * translation of taskfile registers into a |
2576 | * sense descriptors, since that's only |
2577 | * correct for ATA, not ATAPI |
2578 | */ |
2579 | ata_gen_passthru_sense(qc); |
2580 | } |
2581 | |
2582 | /* SCSI EH automatically locks door if sdev->locked is |
2583 | * set. Sometimes door lock request continues to |
2584 | * fail, for example, when no media is present. This |
2585 | * creates a loop - SCSI EH issues door lock which |
2586 | * fails and gets invoked again to acquire sense data |
2587 | * for the failed command. |
2588 | * |
2589 | * If door lock fails, always clear sdev->locked to |
2590 | * avoid this infinite loop. |
2591 | * |
2592 | * This may happen before SCSI scan is complete. Make |
2593 | * sure qc->dev->sdev isn't NULL before dereferencing. |
2594 | */ |
2595 | if (qc->cdb[0] == ALLOW_MEDIUM_REMOVAL && qc->dev->sdev) |
2596 | qc->dev->sdev->locked = 0; |
2597 | |
2598 | qc->scsicmd->result = SAM_STAT_CHECK_CONDITION; |
2599 | qc->scsidone(cmd); |
2600 | ata_qc_free(qc); |
2601 | return; |
2602 | } |
2603 | |
2604 | /* successful completion or old EH failure path */ |
2605 | if (unlikely(err_mask & AC_ERR_DEV)) { |
2606 | cmd->result = SAM_STAT_CHECK_CONDITION; |
2607 | atapi_request_sense(qc); |
2608 | return; |
2609 | } else if (unlikely(err_mask)) { |
2610 | /* FIXME: not quite right; we don't want the |
2611 | * translation of taskfile registers into |
2612 | * a sense descriptors, since that's only |
2613 | * correct for ATA, not ATAPI |
2614 | */ |
2615 | ata_gen_passthru_sense(qc); |
2616 | } else { |
2617 | u8 *scsicmd = cmd->cmnd; |
2618 | |
2619 | if ((scsicmd[0] == INQUIRY) && ((scsicmd[1] & 0x03) == 0)) { |
2620 | unsigned long flags; |
2621 | u8 *buf; |
2622 | |
2623 | buf = ata_scsi_rbuf_get(cmd, true, &flags); |
2624 | |
2625 | /* ATAPI devices typically report zero for their SCSI version, |
2626 | * and sometimes deviate from the spec WRT response data |
2627 | * format. If SCSI version is reported as zero like normal, |
2628 | * then we make the following fixups: 1) Fake MMC-5 version, |
2629 | * to indicate to the Linux scsi midlayer this is a modern |
2630 | * device. 2) Ensure response data format / ATAPI information |
2631 | * are always correct. |
2632 | */ |
2633 | if (buf[2] == 0) { |
2634 | buf[2] = 0x5; |
2635 | buf[3] = 0x32; |
2636 | } |
2637 | |
2638 | ata_scsi_rbuf_put(cmd, true, &flags); |
2639 | } |
2640 | |
2641 | cmd->result = SAM_STAT_GOOD; |
2642 | } |
2643 | |
2644 | qc->scsidone(cmd); |
2645 | ata_qc_free(qc); |
2646 | } |
2647 | /** |
2648 | * atapi_xlat - Initialize PACKET taskfile |
2649 | * @qc: command structure to be initialized |
2650 | * |
2651 | * LOCKING: |
2652 | * spin_lock_irqsave(host lock) |
2653 | * |
2654 | * RETURNS: |
2655 | * Zero on success, non-zero on failure. |
2656 | */ |
2657 | static unsigned int atapi_xlat(struct ata_queued_cmd *qc) |
2658 | { |
2659 | struct scsi_cmnd *scmd = qc->scsicmd; |
2660 | struct ata_device *dev = qc->dev; |
2661 | int nodata = (scmd->sc_data_direction == DMA_NONE); |
2662 | int using_pio = !nodata && (dev->flags & ATA_DFLAG_PIO); |
2663 | unsigned int nbytes; |
2664 | |
2665 | memset(qc->cdb, 0, dev->cdb_len); |
2666 | memcpy(qc->cdb, scmd->cmnd, scmd->cmd_len); |
2667 | |
2668 | qc->complete_fn = atapi_qc_complete; |
2669 | |
2670 | qc->tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
2671 | if (scmd->sc_data_direction == DMA_TO_DEVICE) { |
2672 | qc->tf.flags |= ATA_TFLAG_WRITE; |
2673 | DPRINTK("direction: write\n"); |
2674 | } |
2675 | |
2676 | qc->tf.command = ATA_CMD_PACKET; |
2677 | ata_qc_set_pc_nbytes(qc); |
2678 | |
2679 | /* check whether ATAPI DMA is safe */ |
2680 | if (!nodata && !using_pio && atapi_check_dma(qc)) |
2681 | using_pio = 1; |
2682 | |
2683 | /* Some controller variants snoop this value for Packet |
2684 | * transfers to do state machine and FIFO management. Thus we |
2685 | * want to set it properly, and for DMA where it is |
2686 | * effectively meaningless. |
2687 | */ |
2688 | nbytes = min(ata_qc_raw_nbytes(qc), (unsigned int)63 * 1024); |
2689 | |
2690 | /* Most ATAPI devices which honor transfer chunk size don't |
2691 | * behave according to the spec when odd chunk size which |
2692 | * matches the transfer length is specified. If the number of |
2693 | * bytes to transfer is 2n+1. According to the spec, what |
2694 | * should happen is to indicate that 2n+1 is going to be |
2695 | * transferred and transfer 2n+2 bytes where the last byte is |
2696 | * padding. |
2697 | * |
2698 | * In practice, this doesn't happen. ATAPI devices first |
2699 | * indicate and transfer 2n bytes and then indicate and |
2700 | * transfer 2 bytes where the last byte is padding. |
2701 | * |
2702 | * This inconsistency confuses several controllers which |
2703 | * perform PIO using DMA such as Intel AHCIs and sil3124/32. |
2704 | * These controllers use actual number of transferred bytes to |
2705 | * update DMA poitner and transfer of 4n+2 bytes make those |
2706 | * controller push DMA pointer by 4n+4 bytes because SATA data |
2707 | * FISes are aligned to 4 bytes. This causes data corruption |
2708 | * and buffer overrun. |
2709 | * |
2710 | * Always setting nbytes to even number solves this problem |
2711 | * because then ATAPI devices don't have to split data at 2n |
2712 | * boundaries. |
2713 | */ |
2714 | if (nbytes & 0x1) |
2715 | nbytes++; |
2716 | |
2717 | qc->tf.lbam = (nbytes & 0xFF); |
2718 | qc->tf.lbah = (nbytes >> 8); |
2719 | |
2720 | if (nodata) |
2721 | qc->tf.protocol = ATAPI_PROT_NODATA; |
2722 | else if (using_pio) |
2723 | qc->tf.protocol = ATAPI_PROT_PIO; |
2724 | else { |
2725 | /* DMA data xfer */ |
2726 | qc->tf.protocol = ATAPI_PROT_DMA; |
2727 | qc->tf.feature |= ATAPI_PKT_DMA; |
2728 | |
2729 | if ((dev->flags & ATA_DFLAG_DMADIR) && |
2730 | (scmd->sc_data_direction != DMA_TO_DEVICE)) |
2731 | /* some SATA bridges need us to indicate data xfer direction */ |
2732 | qc->tf.feature |= ATAPI_DMADIR; |
2733 | } |
2734 | |
2735 | |
2736 | /* FIXME: We need to translate 0x05 READ_BLOCK_LIMITS to a MODE_SENSE |
2737 | as ATAPI tape drives don't get this right otherwise */ |
2738 | return 0; |
2739 | } |
2740 | |
2741 | static struct ata_device *ata_find_dev(struct ata_port *ap, int devno) |
2742 | { |
2743 | if (!sata_pmp_attached(ap)) { |
2744 | if (likely(devno < ata_link_max_devices(&ap->link))) |
2745 | return &ap->link.device[devno]; |
2746 | } else { |
2747 | if (likely(devno < ap->nr_pmp_links)) |
2748 | return &ap->pmp_link[devno].device[0]; |
2749 | } |
2750 | |
2751 | return NULL; |
2752 | } |
2753 | |
2754 | static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, |
2755 | const struct scsi_device *scsidev) |
2756 | { |
2757 | int devno; |
2758 | |
2759 | /* skip commands not addressed to targets we simulate */ |
2760 | if (!sata_pmp_attached(ap)) { |
2761 | if (unlikely(scsidev->channel || scsidev->lun)) |
2762 | return NULL; |
2763 | devno = scsidev->id; |
2764 | } else { |
2765 | if (unlikely(scsidev->id || scsidev->lun)) |
2766 | return NULL; |
2767 | devno = scsidev->channel; |
2768 | } |
2769 | |
2770 | return ata_find_dev(ap, devno); |
2771 | } |
2772 | |
2773 | /** |
2774 | * ata_scsi_find_dev - lookup ata_device from scsi_cmnd |
2775 | * @ap: ATA port to which the device is attached |
2776 | * @scsidev: SCSI device from which we derive the ATA device |
2777 | * |
2778 | * Given various information provided in struct scsi_cmnd, |
2779 | * map that onto an ATA bus, and using that mapping |
2780 | * determine which ata_device is associated with the |
2781 | * SCSI command to be sent. |
2782 | * |
2783 | * LOCKING: |
2784 | * spin_lock_irqsave(host lock) |
2785 | * |
2786 | * RETURNS: |
2787 | * Associated ATA device, or %NULL if not found. |
2788 | */ |
2789 | static struct ata_device * |
2790 | ata_scsi_find_dev(struct ata_port *ap, const struct scsi_device *scsidev) |
2791 | { |
2792 | struct ata_device *dev = __ata_scsi_find_dev(ap, scsidev); |
2793 | |
2794 | if (unlikely(!dev || !ata_dev_enabled(dev))) |
2795 | return NULL; |
2796 | |
2797 | return dev; |
2798 | } |
2799 | |
2800 | /* |
2801 | * ata_scsi_map_proto - Map pass-thru protocol value to taskfile value. |
2802 | * @byte1: Byte 1 from pass-thru CDB. |
2803 | * |
2804 | * RETURNS: |
2805 | * ATA_PROT_UNKNOWN if mapping failed/unimplemented, protocol otherwise. |
2806 | */ |
2807 | static u8 |
2808 | ata_scsi_map_proto(u8 byte1) |
2809 | { |
2810 | switch((byte1 & 0x1e) >> 1) { |
2811 | case 3: /* Non-data */ |
2812 | return ATA_PROT_NODATA; |
2813 | |
2814 | case 6: /* DMA */ |
2815 | case 10: /* UDMA Data-in */ |
2816 | case 11: /* UDMA Data-Out */ |
2817 | return ATA_PROT_DMA; |
2818 | |
2819 | case 4: /* PIO Data-in */ |
2820 | case 5: /* PIO Data-out */ |
2821 | return ATA_PROT_PIO; |
2822 | |
2823 | case 0: /* Hard Reset */ |
2824 | case 1: /* SRST */ |
2825 | case 8: /* Device Diagnostic */ |
2826 | case 9: /* Device Reset */ |
2827 | case 7: /* DMA Queued */ |
2828 | case 12: /* FPDMA */ |
2829 | case 15: /* Return Response Info */ |
2830 | default: /* Reserved */ |
2831 | break; |
2832 | } |
2833 | |
2834 | return ATA_PROT_UNKNOWN; |
2835 | } |
2836 | |
2837 | /** |
2838 | * ata_scsi_pass_thru - convert ATA pass-thru CDB to taskfile |
2839 | * @qc: command structure to be initialized |
2840 | * |
2841 | * Handles either 12 or 16-byte versions of the CDB. |
2842 | * |
2843 | * RETURNS: |
2844 | * Zero on success, non-zero on failure. |
2845 | */ |
2846 | static unsigned int ata_scsi_pass_thru(struct ata_queued_cmd *qc) |
2847 | { |
2848 | struct ata_taskfile *tf = &(qc->tf); |
2849 | struct scsi_cmnd *scmd = qc->scsicmd; |
2850 | struct ata_device *dev = qc->dev; |
2851 | const u8 *cdb = scmd->cmnd; |
2852 | |
2853 | if ((tf->protocol = ata_scsi_map_proto(cdb[1])) == ATA_PROT_UNKNOWN) |
2854 | goto invalid_fld; |
2855 | |
2856 | /* |
2857 | * 12 and 16 byte CDBs use different offsets to |
2858 | * provide the various register values. |
2859 | */ |
2860 | if (cdb[0] == ATA_16) { |
2861 | /* |
2862 | * 16-byte CDB - may contain extended commands. |
2863 | * |
2864 | * If that is the case, copy the upper byte register values. |
2865 | */ |
2866 | if (cdb[1] & 0x01) { |
2867 | tf->hob_feature = cdb[3]; |
2868 | tf->hob_nsect = cdb[5]; |
2869 | tf->hob_lbal = cdb[7]; |
2870 | tf->hob_lbam = cdb[9]; |
2871 | tf->hob_lbah = cdb[11]; |
2872 | tf->flags |= ATA_TFLAG_LBA48; |
2873 | } else |
2874 | tf->flags &= ~ATA_TFLAG_LBA48; |
2875 | |
2876 | /* |
2877 | * Always copy low byte, device and command registers. |
2878 | */ |
2879 | tf->feature = cdb[4]; |
2880 | tf->nsect = cdb[6]; |
2881 | tf->lbal = cdb[8]; |
2882 | tf->lbam = cdb[10]; |
2883 | tf->lbah = cdb[12]; |
2884 | tf->device = cdb[13]; |
2885 | tf->command = cdb[14]; |
2886 | } else { |
2887 | /* |
2888 | * 12-byte CDB - incapable of extended commands. |
2889 | */ |
2890 | tf->flags &= ~ATA_TFLAG_LBA48; |
2891 | |
2892 | tf->feature = cdb[3]; |
2893 | tf->nsect = cdb[4]; |
2894 | tf->lbal = cdb[5]; |
2895 | tf->lbam = cdb[6]; |
2896 | tf->lbah = cdb[7]; |
2897 | tf->device = cdb[8]; |
2898 | tf->command = cdb[9]; |
2899 | } |
2900 | |
2901 | /* enforce correct master/slave bit */ |
2902 | tf->device = dev->devno ? |
2903 | tf->device | ATA_DEV1 : tf->device & ~ATA_DEV1; |
2904 | |
2905 | switch (tf->command) { |
2906 | /* READ/WRITE LONG use a non-standard sect_size */ |
2907 | case ATA_CMD_READ_LONG: |
2908 | case ATA_CMD_READ_LONG_ONCE: |
2909 | case ATA_CMD_WRITE_LONG: |
2910 | case ATA_CMD_WRITE_LONG_ONCE: |
2911 | if (tf->protocol != ATA_PROT_PIO || tf->nsect != 1) |
2912 | goto invalid_fld; |
2913 | qc->sect_size = scsi_bufflen(scmd); |
2914 | break; |
2915 | |
2916 | /* commands using reported Logical Block size (e.g. 512 or 4K) */ |
2917 | case ATA_CMD_CFA_WRITE_NE: |
2918 | case ATA_CMD_CFA_TRANS_SECT: |
2919 | case ATA_CMD_CFA_WRITE_MULT_NE: |
2920 | /* XXX: case ATA_CMD_CFA_WRITE_SECTORS_WITHOUT_ERASE: */ |
2921 | case ATA_CMD_READ: |
2922 | case ATA_CMD_READ_EXT: |
2923 | case ATA_CMD_READ_QUEUED: |
2924 | /* XXX: case ATA_CMD_READ_QUEUED_EXT: */ |
2925 | case ATA_CMD_FPDMA_READ: |
2926 | case ATA_CMD_READ_MULTI: |
2927 | case ATA_CMD_READ_MULTI_EXT: |
2928 | case ATA_CMD_PIO_READ: |
2929 | case ATA_CMD_PIO_READ_EXT: |
2930 | case ATA_CMD_READ_STREAM_DMA_EXT: |
2931 | case ATA_CMD_READ_STREAM_EXT: |
2932 | case ATA_CMD_VERIFY: |
2933 | case ATA_CMD_VERIFY_EXT: |
2934 | case ATA_CMD_WRITE: |
2935 | case ATA_CMD_WRITE_EXT: |
2936 | case ATA_CMD_WRITE_FUA_EXT: |
2937 | case ATA_CMD_WRITE_QUEUED: |
2938 | case ATA_CMD_WRITE_QUEUED_FUA_EXT: |
2939 | case ATA_CMD_FPDMA_WRITE: |
2940 | case ATA_CMD_WRITE_MULTI: |
2941 | case ATA_CMD_WRITE_MULTI_EXT: |
2942 | case ATA_CMD_WRITE_MULTI_FUA_EXT: |
2943 | case ATA_CMD_PIO_WRITE: |
2944 | case ATA_CMD_PIO_WRITE_EXT: |
2945 | case ATA_CMD_WRITE_STREAM_DMA_EXT: |
2946 | case ATA_CMD_WRITE_STREAM_EXT: |
2947 | qc->sect_size = scmd->device->sector_size; |
2948 | break; |
2949 | |
2950 | /* Everything else uses 512 byte "sectors" */ |
2951 | default: |
2952 | qc->sect_size = ATA_SECT_SIZE; |
2953 | } |
2954 | |
2955 | /* |
2956 | * Set flags so that all registers will be written, pass on |
2957 | * write indication (used for PIO/DMA setup), result TF is |
2958 | * copied back and we don't whine too much about its failure. |
2959 | */ |
2960 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; |
2961 | if (scmd->sc_data_direction == DMA_TO_DEVICE) |
2962 | tf->flags |= ATA_TFLAG_WRITE; |
2963 | |
2964 | qc->flags |= ATA_QCFLAG_RESULT_TF | ATA_QCFLAG_QUIET; |
2965 | |
2966 | /* |
2967 | * Set transfer length. |
2968 | * |
2969 | * TODO: find out if we need to do more here to |
2970 | * cover scatter/gather case. |
2971 | */ |
2972 | ata_qc_set_pc_nbytes(qc); |
2973 | |
2974 | /* We may not issue DMA commands if no DMA mode is set */ |
2975 | if (tf->protocol == ATA_PROT_DMA && dev->dma_mode == 0) |
2976 | goto invalid_fld; |
2977 | |
2978 | /* sanity check for pio multi commands */ |
2979 | if ((cdb[1] & 0xe0) && !is_multi_taskfile(tf)) |
2980 | goto invalid_fld; |
2981 | |
2982 | if (is_multi_taskfile(tf)) { |
2983 | unsigned int multi_count = 1 << (cdb[1] >> 5); |
2984 | |
2985 | /* compare the passed through multi_count |
2986 | * with the cached multi_count of libata |
2987 | */ |
2988 | if (multi_count != dev->multi_count) |
2989 | ata_dev_warn(dev, "invalid multi_count %u ignored\n", |
2990 | multi_count); |
2991 | } |
2992 | |
2993 | /* |
2994 | * Filter SET_FEATURES - XFER MODE command -- otherwise, |
2995 | * SET_FEATURES - XFER MODE must be preceded/succeeded |
2996 | * by an update to hardware-specific registers for each |
2997 | * controller (i.e. the reason for ->set_piomode(), |
2998 | * ->set_dmamode(), and ->post_set_mode() hooks). |
2999 | */ |
3000 | if (tf->command == ATA_CMD_SET_FEATURES && |
3001 | tf->feature == SETFEATURES_XFER) |
3002 | goto invalid_fld; |
3003 | |
3004 | /* |
3005 | * Filter TPM commands by default. These provide an |
3006 | * essentially uncontrolled encrypted "back door" between |
3007 | * applications and the disk. Set libata.allow_tpm=1 if you |
3008 | * have a real reason for wanting to use them. This ensures |
3009 | * that installed software cannot easily mess stuff up without |
3010 | * user intent. DVR type users will probably ship with this enabled |
3011 | * for movie content management. |
3012 | * |
3013 | * Note that for ATA8 we can issue a DCS change and DCS freeze lock |
3014 | * for this and should do in future but that it is not sufficient as |
3015 | * DCS is an optional feature set. Thus we also do the software filter |
3016 | * so that we comply with the TC consortium stated goal that the user |
3017 | * can turn off TC features of their system. |
3018 | */ |
3019 | if (tf->command >= 0x5C && tf->command <= 0x5F && !libata_allow_tpm) |
3020 | goto invalid_fld; |
3021 | |
3022 | return 0; |
3023 | |
3024 | invalid_fld: |
3025 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); |
3026 | /* "Invalid field in cdb" */ |
3027 | return 1; |
3028 | } |
3029 | |
3030 | static unsigned int ata_scsi_write_same_xlat(struct ata_queued_cmd *qc) |
3031 | { |
3032 | struct ata_taskfile *tf = &qc->tf; |
3033 | struct scsi_cmnd *scmd = qc->scsicmd; |
3034 | struct ata_device *dev = qc->dev; |
3035 | const u8 *cdb = scmd->cmnd; |
3036 | u64 block; |
3037 | u32 n_block; |
3038 | u32 size; |
3039 | void *buf; |
3040 | |
3041 | /* we may not issue DMA commands if no DMA mode is set */ |
3042 | if (unlikely(!dev->dma_mode)) |
3043 | goto invalid_fld; |
3044 | |
3045 | if (unlikely(scmd->cmd_len < 16)) |
3046 | goto invalid_fld; |
3047 | scsi_16_lba_len(cdb, &block, &n_block); |
3048 | |
3049 | /* for now we only support WRITE SAME with the unmap bit set */ |
3050 | if (unlikely(!(cdb[1] & 0x8))) |
3051 | goto invalid_fld; |
3052 | |
3053 | /* |
3054 | * WRITE SAME always has a sector sized buffer as payload, this |
3055 | * should never be a multiple entry S/G list. |
3056 | */ |
3057 | if (!scsi_sg_count(scmd)) |
3058 | goto invalid_fld; |
3059 | |
3060 | buf = page_address(sg_page(scsi_sglist(scmd))); |
3061 | size = ata_set_lba_range_entries(buf, 512, block, n_block); |
3062 | |
3063 | tf->protocol = ATA_PROT_DMA; |
3064 | tf->hob_feature = 0; |
3065 | tf->feature = ATA_DSM_TRIM; |
3066 | tf->hob_nsect = (size / 512) >> 8; |
3067 | tf->nsect = size / 512; |
3068 | tf->command = ATA_CMD_DSM; |
3069 | tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE | ATA_TFLAG_LBA48 | |
3070 | ATA_TFLAG_WRITE; |
3071 | |
3072 | ata_qc_set_pc_nbytes(qc); |
3073 | |
3074 | return 0; |
3075 | |
3076 | invalid_fld: |
3077 | ata_scsi_set_sense(scmd, ILLEGAL_REQUEST, 0x24, 0x00); |
3078 | /* "Invalid field in cdb" */ |
3079 | return 1; |
3080 | } |
3081 | |
3082 | /** |
3083 | * ata_get_xlat_func - check if SCSI to ATA translation is possible |
3084 | * @dev: ATA device |
3085 | * @cmd: SCSI command opcode to consider |
3086 | * |
3087 | * Look up the SCSI command given, and determine whether the |
3088 | * SCSI command is to be translated or simulated. |
3089 | * |
3090 | * RETURNS: |
3091 | * Pointer to translation function if possible, %NULL if not. |
3092 | */ |
3093 | |
3094 | static inline ata_xlat_func_t ata_get_xlat_func(struct ata_device *dev, u8 cmd) |
3095 | { |
3096 | switch (cmd) { |
3097 | case READ_6: |
3098 | case READ_10: |
3099 | case READ_16: |
3100 | |
3101 | case WRITE_6: |
3102 | case WRITE_10: |
3103 | case WRITE_16: |
3104 | return ata_scsi_rw_xlat; |
3105 | |
3106 | case WRITE_SAME_16: |
3107 | return ata_scsi_write_same_xlat; |
3108 | |
3109 | case SYNCHRONIZE_CACHE: |
3110 | if (ata_try_flush_cache(dev)) |
3111 | return ata_scsi_flush_xlat; |
3112 | break; |
3113 | |
3114 | case VERIFY: |
3115 | case VERIFY_16: |
3116 | return ata_scsi_verify_xlat; |
3117 | |
3118 | case ATA_12: |
3119 | case ATA_16: |
3120 | return ata_scsi_pass_thru; |
3121 | |
3122 | case START_STOP: |
3123 | return ata_scsi_start_stop_xlat; |
3124 | } |
3125 | |
3126 | return NULL; |
3127 | } |
3128 | |
3129 | /** |
3130 | * ata_scsi_dump_cdb - dump SCSI command contents to dmesg |
3131 | * @ap: ATA port to which the command was being sent |
3132 | * @cmd: SCSI command to dump |
3133 | * |
3134 | * Prints the contents of a SCSI command via printk(). |
3135 | */ |
3136 | |
3137 | static inline void ata_scsi_dump_cdb(struct ata_port *ap, |
3138 | struct scsi_cmnd *cmd) |
3139 | { |
3140 | #ifdef ATA_DEBUG |
3141 | struct scsi_device *scsidev = cmd->device; |
3142 | u8 *scsicmd = cmd->cmnd; |
3143 | |
3144 | DPRINTK("CDB (%u:%d,%d,%d) %02x %02x %02x %02x %02x %02x %02x %02x %02x\n", |
3145 | ap->print_id, |
3146 | scsidev->channel, scsidev->id, scsidev->lun, |
3147 | scsicmd[0], scsicmd[1], scsicmd[2], scsicmd[3], |
3148 | scsicmd[4], scsicmd[5], scsicmd[6], scsicmd[7], |
3149 | scsicmd[8]); |
3150 | #endif |
3151 | } |
3152 | |
3153 | static inline int __ata_scsi_queuecmd(struct scsi_cmnd *scmd, |
3154 | struct ata_device *dev) |
3155 | { |
3156 | u8 scsi_op = scmd->cmnd[0]; |
3157 | ata_xlat_func_t xlat_func; |
3158 | int rc = 0; |
3159 | |
3160 | if (dev->class == ATA_DEV_ATA) { |
3161 | if (unlikely(!scmd->cmd_len || scmd->cmd_len > dev->cdb_len)) |
3162 | goto bad_cdb_len; |
3163 | |
3164 | xlat_func = ata_get_xlat_func(dev, scsi_op); |
3165 | } else { |
3166 | if (unlikely(!scmd->cmd_len)) |
3167 | goto bad_cdb_len; |
3168 | |
3169 | xlat_func = NULL; |
3170 | if (likely((scsi_op != ATA_16) || !atapi_passthru16)) { |
3171 | /* relay SCSI command to ATAPI device */ |
3172 | int len = COMMAND_SIZE(scsi_op); |
3173 | if (unlikely(len > scmd->cmd_len || len > dev->cdb_len)) |
3174 | goto bad_cdb_len; |
3175 | |
3176 | xlat_func = atapi_xlat; |
3177 | } else { |
3178 | /* ATA_16 passthru, treat as an ATA command */ |
3179 | if (unlikely(scmd->cmd_len > 16)) |
3180 | goto bad_cdb_len; |
3181 | |
3182 | xlat_func = ata_get_xlat_func(dev, scsi_op); |
3183 | } |
3184 | } |
3185 | |
3186 | if (xlat_func) |
3187 | rc = ata_scsi_translate(dev, scmd, xlat_func); |
3188 | else |
3189 | ata_scsi_simulate(dev, scmd); |
3190 | |
3191 | return rc; |
3192 | |
3193 | bad_cdb_len: |
3194 | DPRINTK("bad CDB len=%u, scsi_op=0x%02x, max=%u\n", |
3195 | scmd->cmd_len, scsi_op, dev->cdb_len); |
3196 | scmd->result = DID_ERROR << 16; |
3197 | scmd->scsi_done(scmd); |
3198 | return 0; |
3199 | } |
3200 | |
3201 | /** |
3202 | * ata_scsi_queuecmd - Issue SCSI cdb to libata-managed device |
3203 | * @shost: SCSI host of command to be sent |
3204 | * @cmd: SCSI command to be sent |
3205 | * |
3206 | * In some cases, this function translates SCSI commands into |
3207 | * ATA taskfiles, and queues the taskfiles to be sent to |
3208 | * hardware. In other cases, this function simulates a |
3209 | * SCSI device by evaluating and responding to certain |
3210 | * SCSI commands. This creates the overall effect of |
3211 | * ATA and ATAPI devices appearing as SCSI devices. |
3212 | * |
3213 | * LOCKING: |
3214 | * ATA host lock |
3215 | * |
3216 | * RETURNS: |
3217 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, |
3218 | * 0 otherwise. |
3219 | */ |
3220 | int ata_scsi_queuecmd(struct Scsi_Host *shost, struct scsi_cmnd *cmd) |
3221 | { |
3222 | struct ata_port *ap; |
3223 | struct ata_device *dev; |
3224 | struct scsi_device *scsidev = cmd->device; |
3225 | int rc = 0; |
3226 | unsigned long irq_flags; |
3227 | |
3228 | ap = ata_shost_to_port(shost); |
3229 | |
3230 | spin_lock_irqsave(ap->lock, irq_flags); |
3231 | |
3232 | ata_scsi_dump_cdb(ap, cmd); |
3233 | |
3234 | dev = ata_scsi_find_dev(ap, scsidev); |
3235 | if (likely(dev)) |
3236 | rc = __ata_scsi_queuecmd(cmd, dev); |
3237 | else { |
3238 | cmd->result = (DID_BAD_TARGET << 16); |
3239 | cmd->scsi_done(cmd); |
3240 | } |
3241 | |
3242 | spin_unlock_irqrestore(ap->lock, irq_flags); |
3243 | |
3244 | return rc; |
3245 | } |
3246 | |
3247 | /** |
3248 | * ata_scsi_simulate - simulate SCSI command on ATA device |
3249 | * @dev: the target device |
3250 | * @cmd: SCSI command being sent to device. |
3251 | * |
3252 | * Interprets and directly executes a select list of SCSI commands |
3253 | * that can be handled internally. |
3254 | * |
3255 | * LOCKING: |
3256 | * spin_lock_irqsave(host lock) |
3257 | */ |
3258 | |
3259 | void ata_scsi_simulate(struct ata_device *dev, struct scsi_cmnd *cmd) |
3260 | { |
3261 | struct ata_scsi_args args; |
3262 | const u8 *scsicmd = cmd->cmnd; |
3263 | u8 tmp8; |
3264 | |
3265 | args.dev = dev; |
3266 | args.id = dev->id; |
3267 | args.cmd = cmd; |
3268 | args.done = cmd->scsi_done; |
3269 | |
3270 | switch(scsicmd[0]) { |
3271 | /* TODO: worth improving? */ |
3272 | case FORMAT_UNIT: |
3273 | ata_scsi_invalid_field(cmd); |
3274 | break; |
3275 | |
3276 | case INQUIRY: |
3277 | if (scsicmd[1] & 2) /* is CmdDt set? */ |
3278 | ata_scsi_invalid_field(cmd); |
3279 | else if ((scsicmd[1] & 1) == 0) /* is EVPD clear? */ |
3280 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_std); |
3281 | else switch (scsicmd[2]) { |
3282 | case 0x00: |
3283 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_00); |
3284 | break; |
3285 | case 0x80: |
3286 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_80); |
3287 | break; |
3288 | case 0x83: |
3289 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_83); |
3290 | break; |
3291 | case 0x89: |
3292 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_89); |
3293 | break; |
3294 | case 0xb0: |
3295 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b0); |
3296 | break; |
3297 | case 0xb1: |
3298 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b1); |
3299 | break; |
3300 | case 0xb2: |
3301 | ata_scsi_rbuf_fill(&args, ata_scsiop_inq_b2); |
3302 | break; |
3303 | default: |
3304 | ata_scsi_invalid_field(cmd); |
3305 | break; |
3306 | } |
3307 | break; |
3308 | |
3309 | case MODE_SENSE: |
3310 | case MODE_SENSE_10: |
3311 | ata_scsi_rbuf_fill(&args, ata_scsiop_mode_sense); |
3312 | break; |
3313 | |
3314 | case MODE_SELECT: /* unconditionally return */ |
3315 | case MODE_SELECT_10: /* bad-field-in-cdb */ |
3316 | ata_scsi_invalid_field(cmd); |
3317 | break; |
3318 | |
3319 | case READ_CAPACITY: |
3320 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); |
3321 | break; |
3322 | |
3323 | case SERVICE_ACTION_IN: |
3324 | if ((scsicmd[1] & 0x1f) == SAI_READ_CAPACITY_16) |
3325 | ata_scsi_rbuf_fill(&args, ata_scsiop_read_cap); |
3326 | else |
3327 | ata_scsi_invalid_field(cmd); |
3328 | break; |
3329 | |
3330 | case REPORT_LUNS: |
3331 | ata_scsi_rbuf_fill(&args, ata_scsiop_report_luns); |
3332 | break; |
3333 | |
3334 | case REQUEST_SENSE: |
3335 | ata_scsi_set_sense(cmd, 0, 0, 0); |
3336 | cmd->result = (DRIVER_SENSE << 24); |
3337 | cmd->scsi_done(cmd); |
3338 | break; |
3339 | |
3340 | /* if we reach this, then writeback caching is disabled, |
3341 | * turning this into a no-op. |
3342 | */ |
3343 | case SYNCHRONIZE_CACHE: |
3344 | /* fall through */ |
3345 | |
3346 | /* no-op's, complete with success */ |
3347 | case REZERO_UNIT: |
3348 | case SEEK_6: |
3349 | case SEEK_10: |
3350 | case TEST_UNIT_READY: |
3351 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); |
3352 | break; |
3353 | |
3354 | case SEND_DIAGNOSTIC: |
3355 | tmp8 = scsicmd[1] & ~(1 << 3); |
3356 | if ((tmp8 == 0x4) && (!scsicmd[3]) && (!scsicmd[4])) |
3357 | ata_scsi_rbuf_fill(&args, ata_scsiop_noop); |
3358 | else |
3359 | ata_scsi_invalid_field(cmd); |
3360 | break; |
3361 | |
3362 | /* all other commands */ |
3363 | default: |
3364 | ata_scsi_set_sense(cmd, ILLEGAL_REQUEST, 0x20, 0x0); |
3365 | /* "Invalid command operation code" */ |
3366 | cmd->scsi_done(cmd); |
3367 | break; |
3368 | } |
3369 | } |
3370 | |
3371 | int ata_scsi_add_hosts(struct ata_host *host, struct scsi_host_template *sht) |
3372 | { |
3373 | int i, rc; |
3374 | |
3375 | for (i = 0; i < host->n_ports; i++) { |
3376 | struct ata_port *ap = host->ports[i]; |
3377 | struct Scsi_Host *shost; |
3378 | |
3379 | rc = -ENOMEM; |
3380 | shost = scsi_host_alloc(sht, sizeof(struct ata_port *)); |
3381 | if (!shost) |
3382 | goto err_alloc; |
3383 | |
3384 | shost->eh_noresume = 1; |
3385 | *(struct ata_port **)&shost->hostdata[0] = ap; |
3386 | ap->scsi_host = shost; |
3387 | |
3388 | shost->transportt = ata_scsi_transport_template; |
3389 | shost->unique_id = ap->print_id; |
3390 | shost->max_id = 16; |
3391 | shost->max_lun = 1; |
3392 | shost->max_channel = 1; |
3393 | shost->max_cmd_len = 16; |
3394 | |
3395 | /* Schedule policy is determined by ->qc_defer() |
3396 | * callback and it needs to see every deferred qc. |
3397 | * Set host_blocked to 1 to prevent SCSI midlayer from |
3398 | * automatically deferring requests. |
3399 | */ |
3400 | shost->max_host_blocked = 1; |
3401 | |
3402 | rc = scsi_add_host_with_dma(ap->scsi_host, |
3403 | &ap->tdev, ap->host->dev); |
3404 | if (rc) |
3405 | goto err_add; |
3406 | } |
3407 | |
3408 | return 0; |
3409 | |
3410 | err_add: |
3411 | scsi_host_put(host->ports[i]->scsi_host); |
3412 | err_alloc: |
3413 | while (--i >= 0) { |
3414 | struct Scsi_Host *shost = host->ports[i]->scsi_host; |
3415 | |
3416 | scsi_remove_host(shost); |
3417 | scsi_host_put(shost); |
3418 | } |
3419 | return rc; |
3420 | } |
3421 | |
3422 | void ata_scsi_scan_host(struct ata_port *ap, int sync) |
3423 | { |
3424 | int tries = 5; |
3425 | struct ata_device *last_failed_dev = NULL; |
3426 | struct ata_link *link; |
3427 | struct ata_device *dev; |
3428 | |
3429 | repeat: |
3430 | ata_for_each_link(link, ap, EDGE) { |
3431 | ata_for_each_dev(dev, link, ENABLED) { |
3432 | struct scsi_device *sdev; |
3433 | int channel = 0, id = 0; |
3434 | |
3435 | if (dev->sdev) |
3436 | continue; |
3437 | |
3438 | if (ata_is_host_link(link)) |
3439 | id = dev->devno; |
3440 | else |
3441 | channel = link->pmp; |
3442 | |
3443 | sdev = __scsi_add_device(ap->scsi_host, channel, id, 0, |
3444 | NULL); |
3445 | if (!IS_ERR(sdev)) { |
3446 | dev->sdev = sdev; |
3447 | scsi_device_put(sdev); |
3448 | ata_acpi_bind(dev); |
3449 | } else { |
3450 | dev->sdev = NULL; |
3451 | } |
3452 | } |
3453 | } |
3454 | |
3455 | /* If we scanned while EH was in progress or allocation |
3456 | * failure occurred, scan would have failed silently. Check |
3457 | * whether all devices are attached. |
3458 | */ |
3459 | ata_for_each_link(link, ap, EDGE) { |
3460 | ata_for_each_dev(dev, link, ENABLED) { |
3461 | if (!dev->sdev) |
3462 | goto exit_loop; |
3463 | } |
3464 | } |
3465 | exit_loop: |
3466 | if (!link) |
3467 | return; |
3468 | |
3469 | /* we're missing some SCSI devices */ |
3470 | if (sync) { |
3471 | /* If caller requested synchrnous scan && we've made |
3472 | * any progress, sleep briefly and repeat. |
3473 | */ |
3474 | if (dev != last_failed_dev) { |
3475 | msleep(100); |
3476 | last_failed_dev = dev; |
3477 | goto repeat; |
3478 | } |
3479 | |
3480 | /* We might be failing to detect boot device, give it |
3481 | * a few more chances. |
3482 | */ |
3483 | if (--tries) { |
3484 | msleep(100); |
3485 | goto repeat; |
3486 | } |
3487 | |
3488 | ata_port_err(ap, |
3489 | "WARNING: synchronous SCSI scan failed without making any progress, switching to async\n"); |
3490 | } |
3491 | |
3492 | queue_delayed_work(system_long_wq, &ap->hotplug_task, |
3493 | round_jiffies_relative(HZ)); |
3494 | } |
3495 | |
3496 | /** |
3497 | * ata_scsi_offline_dev - offline attached SCSI device |
3498 | * @dev: ATA device to offline attached SCSI device for |
3499 | * |
3500 | * This function is called from ata_eh_hotplug() and responsible |
3501 | * for taking the SCSI device attached to @dev offline. This |
3502 | * function is called with host lock which protects dev->sdev |
3503 | * against clearing. |
3504 | * |
3505 | * LOCKING: |
3506 | * spin_lock_irqsave(host lock) |
3507 | * |
3508 | * RETURNS: |
3509 | * 1 if attached SCSI device exists, 0 otherwise. |
3510 | */ |
3511 | int ata_scsi_offline_dev(struct ata_device *dev) |
3512 | { |
3513 | if (dev->sdev) { |
3514 | scsi_device_set_state(dev->sdev, SDEV_OFFLINE); |
3515 | return 1; |
3516 | } |
3517 | return 0; |
3518 | } |
3519 | |
3520 | /** |
3521 | * ata_scsi_remove_dev - remove attached SCSI device |
3522 | * @dev: ATA device to remove attached SCSI device for |
3523 | * |
3524 | * This function is called from ata_eh_scsi_hotplug() and |
3525 | * responsible for removing the SCSI device attached to @dev. |
3526 | * |
3527 | * LOCKING: |
3528 | * Kernel thread context (may sleep). |
3529 | */ |
3530 | static void ata_scsi_remove_dev(struct ata_device *dev) |
3531 | { |
3532 | struct ata_port *ap = dev->link->ap; |
3533 | struct scsi_device *sdev; |
3534 | unsigned long flags; |
3535 | |
3536 | /* Alas, we need to grab scan_mutex to ensure SCSI device |
3537 | * state doesn't change underneath us and thus |
3538 | * scsi_device_get() always succeeds. The mutex locking can |
3539 | * be removed if there is __scsi_device_get() interface which |
3540 | * increments reference counts regardless of device state. |
3541 | */ |
3542 | mutex_lock(&ap->scsi_host->scan_mutex); |
3543 | spin_lock_irqsave(ap->lock, flags); |
3544 | |
3545 | ata_acpi_unbind(dev); |
3546 | |
3547 | /* clearing dev->sdev is protected by host lock */ |
3548 | sdev = dev->sdev; |
3549 | dev->sdev = NULL; |
3550 | |
3551 | if (sdev) { |
3552 | /* If user initiated unplug races with us, sdev can go |
3553 | * away underneath us after the host lock and |
3554 | * scan_mutex are released. Hold onto it. |
3555 | */ |
3556 | if (scsi_device_get(sdev) == 0) { |
3557 | /* The following ensures the attached sdev is |
3558 | * offline on return from ata_scsi_offline_dev() |
3559 | * regardless it wins or loses the race |
3560 | * against this function. |
3561 | */ |
3562 | scsi_device_set_state(sdev, SDEV_OFFLINE); |
3563 | } else { |
3564 | WARN_ON(1); |
3565 | sdev = NULL; |
3566 | } |
3567 | } |
3568 | |
3569 | spin_unlock_irqrestore(ap->lock, flags); |
3570 | mutex_unlock(&ap->scsi_host->scan_mutex); |
3571 | |
3572 | if (sdev) { |
3573 | ata_dev_info(dev, "detaching (SCSI %s)\n", |
3574 | dev_name(&sdev->sdev_gendev)); |
3575 | |
3576 | scsi_remove_device(sdev); |
3577 | scsi_device_put(sdev); |
3578 | } |
3579 | } |
3580 | |
3581 | static void ata_scsi_handle_link_detach(struct ata_link *link) |
3582 | { |
3583 | struct ata_port *ap = link->ap; |
3584 | struct ata_device *dev; |
3585 | |
3586 | ata_for_each_dev(dev, link, ALL) { |
3587 | unsigned long flags; |
3588 | |
3589 | if (!(dev->flags & ATA_DFLAG_DETACHED)) |
3590 | continue; |
3591 | |
3592 | spin_lock_irqsave(ap->lock, flags); |
3593 | dev->flags &= ~ATA_DFLAG_DETACHED; |
3594 | spin_unlock_irqrestore(ap->lock, flags); |
3595 | |
3596 | ata_scsi_remove_dev(dev); |
3597 | } |
3598 | } |
3599 | |
3600 | /** |
3601 | * ata_scsi_media_change_notify - send media change event |
3602 | * @dev: Pointer to the disk device with media change event |
3603 | * |
3604 | * Tell the block layer to send a media change notification |
3605 | * event. |
3606 | * |
3607 | * LOCKING: |
3608 | * spin_lock_irqsave(host lock) |
3609 | */ |
3610 | void ata_scsi_media_change_notify(struct ata_device *dev) |
3611 | { |
3612 | if (dev->sdev) |
3613 | sdev_evt_send_simple(dev->sdev, SDEV_EVT_MEDIA_CHANGE, |
3614 | GFP_ATOMIC); |
3615 | } |
3616 | |
3617 | /** |
3618 | * ata_scsi_hotplug - SCSI part of hotplug |
3619 | * @work: Pointer to ATA port to perform SCSI hotplug on |
3620 | * |
3621 | * Perform SCSI part of hotplug. It's executed from a separate |
3622 | * workqueue after EH completes. This is necessary because SCSI |
3623 | * hot plugging requires working EH and hot unplugging is |
3624 | * synchronized with hot plugging with a mutex. |
3625 | * |
3626 | * LOCKING: |
3627 | * Kernel thread context (may sleep). |
3628 | */ |
3629 | void ata_scsi_hotplug(struct work_struct *work) |
3630 | { |
3631 | struct ata_port *ap = |
3632 | container_of(work, struct ata_port, hotplug_task.work); |
3633 | int i; |
3634 | |
3635 | if (ap->pflags & ATA_PFLAG_UNLOADING) { |
3636 | DPRINTK("ENTER/EXIT - unloading\n"); |
3637 | return; |
3638 | } |
3639 | |
3640 | DPRINTK("ENTER\n"); |
3641 | mutex_lock(&ap->scsi_scan_mutex); |
3642 | |
3643 | /* Unplug detached devices. We cannot use link iterator here |
3644 | * because PMP links have to be scanned even if PMP is |
3645 | * currently not attached. Iterate manually. |
3646 | */ |
3647 | ata_scsi_handle_link_detach(&ap->link); |
3648 | if (ap->pmp_link) |
3649 | for (i = 0; i < SATA_PMP_MAX_PORTS; i++) |
3650 | ata_scsi_handle_link_detach(&ap->pmp_link[i]); |
3651 | |
3652 | /* scan for new ones */ |
3653 | ata_scsi_scan_host(ap, 0); |
3654 | |
3655 | mutex_unlock(&ap->scsi_scan_mutex); |
3656 | DPRINTK("EXIT\n"); |
3657 | } |
3658 | |
3659 | /** |
3660 | * ata_scsi_user_scan - indication for user-initiated bus scan |
3661 | * @shost: SCSI host to scan |
3662 | * @channel: Channel to scan |
3663 | * @id: ID to scan |
3664 | * @lun: LUN to scan |
3665 | * |
3666 | * This function is called when user explicitly requests bus |
3667 | * scan. Set probe pending flag and invoke EH. |
3668 | * |
3669 | * LOCKING: |
3670 | * SCSI layer (we don't care) |
3671 | * |
3672 | * RETURNS: |
3673 | * Zero. |
3674 | */ |
3675 | int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, |
3676 | unsigned int id, unsigned int lun) |
3677 | { |
3678 | struct ata_port *ap = ata_shost_to_port(shost); |
3679 | unsigned long flags; |
3680 | int devno, rc = 0; |
3681 | |
3682 | if (!ap->ops->error_handler) |
3683 | return -EOPNOTSUPP; |
3684 | |
3685 | if (lun != SCAN_WILD_CARD && lun) |
3686 | return -EINVAL; |
3687 | |
3688 | if (!sata_pmp_attached(ap)) { |
3689 | if (channel != SCAN_WILD_CARD && channel) |
3690 | return -EINVAL; |
3691 | devno = id; |
3692 | } else { |
3693 | if (id != SCAN_WILD_CARD && id) |
3694 | return -EINVAL; |
3695 | devno = channel; |
3696 | } |
3697 | |
3698 | spin_lock_irqsave(ap->lock, flags); |
3699 | |
3700 | if (devno == SCAN_WILD_CARD) { |
3701 | struct ata_link *link; |
3702 | |
3703 | ata_for_each_link(link, ap, EDGE) { |
3704 | struct ata_eh_info *ehi = &link->eh_info; |
3705 | ehi->probe_mask |= ATA_ALL_DEVICES; |
3706 | ehi->action |= ATA_EH_RESET; |
3707 | } |
3708 | } else { |
3709 | struct ata_device *dev = ata_find_dev(ap, devno); |
3710 | |
3711 | if (dev) { |
3712 | struct ata_eh_info *ehi = &dev->link->eh_info; |
3713 | ehi->probe_mask |= 1 << dev->devno; |
3714 | ehi->action |= ATA_EH_RESET; |
3715 | } else |
3716 | rc = -EINVAL; |
3717 | } |
3718 | |
3719 | if (rc == 0) { |
3720 | ata_port_schedule_eh(ap); |
3721 | spin_unlock_irqrestore(ap->lock, flags); |
3722 | ata_port_wait_eh(ap); |
3723 | } else |
3724 | spin_unlock_irqrestore(ap->lock, flags); |
3725 | |
3726 | return rc; |
3727 | } |
3728 | |
3729 | /** |
3730 | * ata_scsi_dev_rescan - initiate scsi_rescan_device() |
3731 | * @work: Pointer to ATA port to perform scsi_rescan_device() |
3732 | * |
3733 | * After ATA pass thru (SAT) commands are executed successfully, |
3734 | * libata need to propagate the changes to SCSI layer. |
3735 | * |
3736 | * LOCKING: |
3737 | * Kernel thread context (may sleep). |
3738 | */ |
3739 | void ata_scsi_dev_rescan(struct work_struct *work) |
3740 | { |
3741 | struct ata_port *ap = |
3742 | container_of(work, struct ata_port, scsi_rescan_task); |
3743 | struct ata_link *link; |
3744 | struct ata_device *dev; |
3745 | unsigned long flags; |
3746 | |
3747 | mutex_lock(&ap->scsi_scan_mutex); |
3748 | spin_lock_irqsave(ap->lock, flags); |
3749 | |
3750 | ata_for_each_link(link, ap, EDGE) { |
3751 | ata_for_each_dev(dev, link, ENABLED) { |
3752 | struct scsi_device *sdev = dev->sdev; |
3753 | |
3754 | if (!sdev) |
3755 | continue; |
3756 | if (scsi_device_get(sdev)) |
3757 | continue; |
3758 | |
3759 | spin_unlock_irqrestore(ap->lock, flags); |
3760 | scsi_rescan_device(&(sdev->sdev_gendev)); |
3761 | scsi_device_put(sdev); |
3762 | spin_lock_irqsave(ap->lock, flags); |
3763 | } |
3764 | } |
3765 | |
3766 | spin_unlock_irqrestore(ap->lock, flags); |
3767 | mutex_unlock(&ap->scsi_scan_mutex); |
3768 | } |
3769 | |
3770 | /** |
3771 | * ata_sas_port_alloc - Allocate port for a SAS attached SATA device |
3772 | * @host: ATA host container for all SAS ports |
3773 | * @port_info: Information from low-level host driver |
3774 | * @shost: SCSI host that the scsi device is attached to |
3775 | * |
3776 | * LOCKING: |
3777 | * PCI/etc. bus probe sem. |
3778 | * |
3779 | * RETURNS: |
3780 | * ata_port pointer on success / NULL on failure. |
3781 | */ |
3782 | |
3783 | struct ata_port *ata_sas_port_alloc(struct ata_host *host, |
3784 | struct ata_port_info *port_info, |
3785 | struct Scsi_Host *shost) |
3786 | { |
3787 | struct ata_port *ap; |
3788 | |
3789 | ap = ata_port_alloc(host); |
3790 | if (!ap) |
3791 | return NULL; |
3792 | |
3793 | ap->port_no = 0; |
3794 | ap->lock = &host->lock; |
3795 | ap->pio_mask = port_info->pio_mask; |
3796 | ap->mwdma_mask = port_info->mwdma_mask; |
3797 | ap->udma_mask = port_info->udma_mask; |
3798 | ap->flags |= port_info->flags; |
3799 | ap->ops = port_info->port_ops; |
3800 | ap->cbl = ATA_CBL_SATA; |
3801 | |
3802 | return ap; |
3803 | } |
3804 | EXPORT_SYMBOL_GPL(ata_sas_port_alloc); |
3805 | |
3806 | /** |
3807 | * ata_sas_port_start - Set port up for dma. |
3808 | * @ap: Port to initialize |
3809 | * |
3810 | * Called just after data structures for each port are |
3811 | * initialized. |
3812 | * |
3813 | * May be used as the port_start() entry in ata_port_operations. |
3814 | * |
3815 | * LOCKING: |
3816 | * Inherited from caller. |
3817 | */ |
3818 | int ata_sas_port_start(struct ata_port *ap) |
3819 | { |
3820 | /* |
3821 | * the port is marked as frozen at allocation time, but if we don't |
3822 | * have new eh, we won't thaw it |
3823 | */ |
3824 | if (!ap->ops->error_handler) |
3825 | ap->pflags &= ~ATA_PFLAG_FROZEN; |
3826 | return 0; |
3827 | } |
3828 | EXPORT_SYMBOL_GPL(ata_sas_port_start); |
3829 | |
3830 | /** |
3831 | * ata_port_stop - Undo ata_sas_port_start() |
3832 | * @ap: Port to shut down |
3833 | * |
3834 | * May be used as the port_stop() entry in ata_port_operations. |
3835 | * |
3836 | * LOCKING: |
3837 | * Inherited from caller. |
3838 | */ |
3839 | |
3840 | void ata_sas_port_stop(struct ata_port *ap) |
3841 | { |
3842 | } |
3843 | EXPORT_SYMBOL_GPL(ata_sas_port_stop); |
3844 | |
3845 | /** |
3846 | * ata_sas_async_probe - simply schedule probing and return |
3847 | * @ap: Port to probe |
3848 | * |
3849 | * For batch scheduling of probe for sas attached ata devices, assumes |
3850 | * the port has already been through ata_sas_port_init() |
3851 | */ |
3852 | void ata_sas_async_probe(struct ata_port *ap) |
3853 | { |
3854 | __ata_port_probe(ap); |
3855 | } |
3856 | EXPORT_SYMBOL_GPL(ata_sas_async_probe); |
3857 | |
3858 | int ata_sas_sync_probe(struct ata_port *ap) |
3859 | { |
3860 | return ata_port_probe(ap); |
3861 | } |
3862 | EXPORT_SYMBOL_GPL(ata_sas_sync_probe); |
3863 | |
3864 | |
3865 | /** |
3866 | * ata_sas_port_init - Initialize a SATA device |
3867 | * @ap: SATA port to initialize |
3868 | * |
3869 | * LOCKING: |
3870 | * PCI/etc. bus probe sem. |
3871 | * |
3872 | * RETURNS: |
3873 | * Zero on success, non-zero on error. |
3874 | */ |
3875 | |
3876 | int ata_sas_port_init(struct ata_port *ap) |
3877 | { |
3878 | int rc = ap->ops->port_start(ap); |
3879 | |
3880 | if (rc) |
3881 | return rc; |
3882 | ap->print_id = atomic_inc_return(&ata_print_id); |
3883 | return 0; |
3884 | } |
3885 | EXPORT_SYMBOL_GPL(ata_sas_port_init); |
3886 | |
3887 | /** |
3888 | * ata_sas_port_destroy - Destroy a SATA port allocated by ata_sas_port_alloc |
3889 | * @ap: SATA port to destroy |
3890 | * |
3891 | */ |
3892 | |
3893 | void ata_sas_port_destroy(struct ata_port *ap) |
3894 | { |
3895 | if (ap->ops->port_stop) |
3896 | ap->ops->port_stop(ap); |
3897 | kfree(ap); |
3898 | } |
3899 | EXPORT_SYMBOL_GPL(ata_sas_port_destroy); |
3900 | |
3901 | /** |
3902 | * ata_sas_slave_configure - Default slave_config routine for libata devices |
3903 | * @sdev: SCSI device to configure |
3904 | * @ap: ATA port to which SCSI device is attached |
3905 | * |
3906 | * RETURNS: |
3907 | * Zero. |
3908 | */ |
3909 | |
3910 | int ata_sas_slave_configure(struct scsi_device *sdev, struct ata_port *ap) |
3911 | { |
3912 | ata_scsi_sdev_config(sdev); |
3913 | ata_scsi_dev_config(sdev, ap->link.device); |
3914 | return 0; |
3915 | } |
3916 | EXPORT_SYMBOL_GPL(ata_sas_slave_configure); |
3917 | |
3918 | /** |
3919 | * ata_sas_queuecmd - Issue SCSI cdb to libata-managed device |
3920 | * @cmd: SCSI command to be sent |
3921 | * @ap: ATA port to which the command is being sent |
3922 | * |
3923 | * RETURNS: |
3924 | * Return value from __ata_scsi_queuecmd() if @cmd can be queued, |
3925 | * 0 otherwise. |
3926 | */ |
3927 | |
3928 | int ata_sas_queuecmd(struct scsi_cmnd *cmd, struct ata_port *ap) |
3929 | { |
3930 | int rc = 0; |
3931 | |
3932 | ata_scsi_dump_cdb(ap, cmd); |
3933 | |
3934 | if (likely(ata_dev_enabled(ap->link.device))) |
3935 | rc = __ata_scsi_queuecmd(cmd, ap->link.device); |
3936 | else { |
3937 | cmd->result = (DID_BAD_TARGET << 16); |
3938 | cmd->scsi_done(cmd); |
3939 | } |
3940 | return rc; |
3941 | } |
3942 | EXPORT_SYMBOL_GPL(ata_sas_queuecmd); |
3943 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9