Root/
1 | /* |
2 | * EP93XX PATA controller driver. |
3 | * |
4 | * Copyright (c) 2012, Metasoft s.c. |
5 | * Rafal Prylowski <prylowski@metasoft.pl> |
6 | * |
7 | * Based on pata_scc.c, pata_icside.c and on earlier version of EP93XX |
8 | * PATA driver by Lennert Buytenhek and Alessandro Zummo. |
9 | * Read/Write timings, resource management and other improvements |
10 | * from driver by Joao Ramos and Bartlomiej Zolnierkiewicz. |
11 | * DMA engine support based on spi-ep93xx.c by Mika Westerberg. |
12 | * |
13 | * Original copyrights: |
14 | * |
15 | * Support for Cirrus Logic's EP93xx (EP9312, EP9315) CPUs |
16 | * PATA host controller driver. |
17 | * |
18 | * Copyright (c) 2009, Bartlomiej Zolnierkiewicz |
19 | * |
20 | * Heavily based on the ep93xx-ide.c driver: |
21 | * |
22 | * Copyright (c) 2009, Joao Ramos <joao.ramos@inov.pt> |
23 | * INESC Inovacao (INOV) |
24 | * |
25 | * EP93XX PATA controller driver. |
26 | * Copyright (C) 2007 Lennert Buytenhek <buytenh@wantstofly.org> |
27 | * |
28 | * An ATA driver for the Cirrus Logic EP93xx PATA controller. |
29 | * |
30 | * Based on an earlier version by Alessandro Zummo, which is: |
31 | * Copyright (C) 2006 Tower Technologies |
32 | */ |
33 | |
34 | #include <linux/kernel.h> |
35 | #include <linux/module.h> |
36 | #include <linux/init.h> |
37 | #include <linux/blkdev.h> |
38 | #include <scsi/scsi_host.h> |
39 | #include <linux/ata.h> |
40 | #include <linux/libata.h> |
41 | #include <linux/platform_device.h> |
42 | #include <linux/delay.h> |
43 | #include <linux/dmaengine.h> |
44 | #include <linux/ktime.h> |
45 | |
46 | #include <mach/dma.h> |
47 | #include <mach/platform.h> |
48 | |
49 | #define DRV_NAME "ep93xx-ide" |
50 | #define DRV_VERSION "1.0" |
51 | |
52 | enum { |
53 | /* IDE Control Register */ |
54 | IDECTRL = 0x00, |
55 | IDECTRL_CS0N = (1 << 0), |
56 | IDECTRL_CS1N = (1 << 1), |
57 | IDECTRL_DIORN = (1 << 5), |
58 | IDECTRL_DIOWN = (1 << 6), |
59 | IDECTRL_INTRQ = (1 << 9), |
60 | IDECTRL_IORDY = (1 << 10), |
61 | /* |
62 | * the device IDE register to be accessed is selected through |
63 | * IDECTRL register's specific bitfields 'DA', 'CS1N' and 'CS0N': |
64 | * b4 b3 b2 b1 b0 |
65 | * A2 A1 A0 CS1N CS0N |
66 | * the values filled in this structure allows the value to be directly |
67 | * ORed to the IDECTRL register, hence giving directly the A[2:0] and |
68 | * CS1N/CS0N values for each IDE register. |
69 | * The values correspond to the transformation: |
70 | * ((real IDE address) << 2) | CS1N value << 1 | CS0N value |
71 | */ |
72 | IDECTRL_ADDR_CMD = 0 + 2, /* CS1 */ |
73 | IDECTRL_ADDR_DATA = (ATA_REG_DATA << 2) + 2, |
74 | IDECTRL_ADDR_ERROR = (ATA_REG_ERR << 2) + 2, |
75 | IDECTRL_ADDR_FEATURE = (ATA_REG_FEATURE << 2) + 2, |
76 | IDECTRL_ADDR_NSECT = (ATA_REG_NSECT << 2) + 2, |
77 | IDECTRL_ADDR_LBAL = (ATA_REG_LBAL << 2) + 2, |
78 | IDECTRL_ADDR_LBAM = (ATA_REG_LBAM << 2) + 2, |
79 | IDECTRL_ADDR_LBAH = (ATA_REG_LBAH << 2) + 2, |
80 | IDECTRL_ADDR_DEVICE = (ATA_REG_DEVICE << 2) + 2, |
81 | IDECTRL_ADDR_STATUS = (ATA_REG_STATUS << 2) + 2, |
82 | IDECTRL_ADDR_COMMAND = (ATA_REG_CMD << 2) + 2, |
83 | IDECTRL_ADDR_ALTSTATUS = (0x06 << 2) + 1, /* CS0 */ |
84 | IDECTRL_ADDR_CTL = (0x06 << 2) + 1, /* CS0 */ |
85 | |
86 | /* IDE Configuration Register */ |
87 | IDECFG = 0x04, |
88 | IDECFG_IDEEN = (1 << 0), |
89 | IDECFG_PIO = (1 << 1), |
90 | IDECFG_MDMA = (1 << 2), |
91 | IDECFG_UDMA = (1 << 3), |
92 | IDECFG_MODE_SHIFT = 4, |
93 | IDECFG_MODE_MASK = (0xf << 4), |
94 | IDECFG_WST_SHIFT = 8, |
95 | IDECFG_WST_MASK = (0x3 << 8), |
96 | |
97 | /* MDMA Operation Register */ |
98 | IDEMDMAOP = 0x08, |
99 | |
100 | /* UDMA Operation Register */ |
101 | IDEUDMAOP = 0x0c, |
102 | IDEUDMAOP_UEN = (1 << 0), |
103 | IDEUDMAOP_RWOP = (1 << 1), |
104 | |
105 | /* PIO/MDMA/UDMA Data Registers */ |
106 | IDEDATAOUT = 0x10, |
107 | IDEDATAIN = 0x14, |
108 | IDEMDMADATAOUT = 0x18, |
109 | IDEMDMADATAIN = 0x1c, |
110 | IDEUDMADATAOUT = 0x20, |
111 | IDEUDMADATAIN = 0x24, |
112 | |
113 | /* UDMA Status Register */ |
114 | IDEUDMASTS = 0x28, |
115 | IDEUDMASTS_DMAIDE = (1 << 16), |
116 | IDEUDMASTS_INTIDE = (1 << 17), |
117 | IDEUDMASTS_SBUSY = (1 << 18), |
118 | IDEUDMASTS_NDO = (1 << 24), |
119 | IDEUDMASTS_NDI = (1 << 25), |
120 | IDEUDMASTS_N4X = (1 << 26), |
121 | |
122 | /* UDMA Debug Status Register */ |
123 | IDEUDMADEBUG = 0x2c, |
124 | }; |
125 | |
126 | struct ep93xx_pata_data { |
127 | const struct platform_device *pdev; |
128 | void __iomem *ide_base; |
129 | struct ata_timing t; |
130 | bool iordy; |
131 | |
132 | unsigned long udma_in_phys; |
133 | unsigned long udma_out_phys; |
134 | |
135 | struct dma_chan *dma_rx_channel; |
136 | struct ep93xx_dma_data dma_rx_data; |
137 | struct dma_chan *dma_tx_channel; |
138 | struct ep93xx_dma_data dma_tx_data; |
139 | }; |
140 | |
141 | static void ep93xx_pata_clear_regs(void __iomem *base) |
142 | { |
143 | writel(IDECTRL_CS0N | IDECTRL_CS1N | IDECTRL_DIORN | |
144 | IDECTRL_DIOWN, base + IDECTRL); |
145 | |
146 | writel(0, base + IDECFG); |
147 | writel(0, base + IDEMDMAOP); |
148 | writel(0, base + IDEUDMAOP); |
149 | writel(0, base + IDEDATAOUT); |
150 | writel(0, base + IDEDATAIN); |
151 | writel(0, base + IDEMDMADATAOUT); |
152 | writel(0, base + IDEMDMADATAIN); |
153 | writel(0, base + IDEUDMADATAOUT); |
154 | writel(0, base + IDEUDMADATAIN); |
155 | writel(0, base + IDEUDMADEBUG); |
156 | } |
157 | |
158 | static bool ep93xx_pata_check_iordy(void __iomem *base) |
159 | { |
160 | return !!(readl(base + IDECTRL) & IDECTRL_IORDY); |
161 | } |
162 | |
163 | /* |
164 | * According to EP93xx User's Guide, WST field of IDECFG specifies number |
165 | * of HCLK cycles to hold the data bus after a PIO write operation. |
166 | * It should be programmed to guarantee following delays: |
167 | * |
168 | * PIO Mode [ns] |
169 | * 0 30 |
170 | * 1 20 |
171 | * 2 15 |
172 | * 3 10 |
173 | * 4 5 |
174 | * |
175 | * Maximum possible value for HCLK is 100MHz. |
176 | */ |
177 | static int ep93xx_pata_get_wst(int pio_mode) |
178 | { |
179 | int val; |
180 | |
181 | if (pio_mode == 0) |
182 | val = 3; |
183 | else if (pio_mode < 3) |
184 | val = 2; |
185 | else |
186 | val = 1; |
187 | |
188 | return val << IDECFG_WST_SHIFT; |
189 | } |
190 | |
191 | static void ep93xx_pata_enable_pio(void __iomem *base, int pio_mode) |
192 | { |
193 | writel(IDECFG_IDEEN | IDECFG_PIO | |
194 | ep93xx_pata_get_wst(pio_mode) | |
195 | (pio_mode << IDECFG_MODE_SHIFT), base + IDECFG); |
196 | } |
197 | |
198 | /* |
199 | * Based on delay loop found in mach-pxa/mp900.c. |
200 | * |
201 | * Single iteration should take 5 cpu cycles. This is 25ns assuming the |
202 | * fastest ep93xx cpu speed (200MHz) and is better optimized for PIO4 timings |
203 | * than eg. 20ns. |
204 | */ |
205 | static void ep93xx_pata_delay(unsigned long count) |
206 | { |
207 | __asm__ volatile ( |
208 | "0:\n" |
209 | "mov r0, r0\n" |
210 | "subs %0, %1, #1\n" |
211 | "bge 0b\n" |
212 | : "=r" (count) |
213 | : "0" (count) |
214 | ); |
215 | } |
216 | |
217 | static unsigned long ep93xx_pata_wait_for_iordy(void __iomem *base, |
218 | unsigned long t2) |
219 | { |
220 | /* |
221 | * According to ATA specification, IORDY pin can be first sampled |
222 | * tA = 35ns after activation of DIOR-/DIOW-. Maximum IORDY pulse |
223 | * width is tB = 1250ns. |
224 | * |
225 | * We are already t2 delay loop iterations after activation of |
226 | * DIOR-/DIOW-, so we set timeout to (1250 + 35) / 25 - t2 additional |
227 | * delay loop iterations. |
228 | */ |
229 | unsigned long start = (1250 + 35) / 25 - t2; |
230 | unsigned long counter = start; |
231 | |
232 | while (!ep93xx_pata_check_iordy(base) && counter--) |
233 | ep93xx_pata_delay(1); |
234 | return start - counter; |
235 | } |
236 | |
237 | /* common part at start of ep93xx_pata_read/write() */ |
238 | static void ep93xx_pata_rw_begin(void __iomem *base, unsigned long addr, |
239 | unsigned long t1) |
240 | { |
241 | writel(IDECTRL_DIOWN | IDECTRL_DIORN | addr, base + IDECTRL); |
242 | ep93xx_pata_delay(t1); |
243 | } |
244 | |
245 | /* common part at end of ep93xx_pata_read/write() */ |
246 | static void ep93xx_pata_rw_end(void __iomem *base, unsigned long addr, |
247 | bool iordy, unsigned long t0, unsigned long t2, |
248 | unsigned long t2i) |
249 | { |
250 | ep93xx_pata_delay(t2); |
251 | /* lengthen t2 if needed */ |
252 | if (iordy) |
253 | t2 += ep93xx_pata_wait_for_iordy(base, t2); |
254 | writel(IDECTRL_DIOWN | IDECTRL_DIORN | addr, base + IDECTRL); |
255 | if (t0 > t2 && t0 - t2 > t2i) |
256 | ep93xx_pata_delay(t0 - t2); |
257 | else |
258 | ep93xx_pata_delay(t2i); |
259 | } |
260 | |
261 | static u16 ep93xx_pata_read(struct ep93xx_pata_data *drv_data, |
262 | unsigned long addr, |
263 | bool reg) |
264 | { |
265 | void __iomem *base = drv_data->ide_base; |
266 | const struct ata_timing *t = &drv_data->t; |
267 | unsigned long t0 = reg ? t->cyc8b : t->cycle; |
268 | unsigned long t2 = reg ? t->act8b : t->active; |
269 | unsigned long t2i = reg ? t->rec8b : t->recover; |
270 | |
271 | ep93xx_pata_rw_begin(base, addr, t->setup); |
272 | writel(IDECTRL_DIOWN | addr, base + IDECTRL); |
273 | /* |
274 | * The IDEDATAIN register is loaded from the DD pins at the positive |
275 | * edge of the DIORN signal. (EP93xx UG p27-14) |
276 | */ |
277 | ep93xx_pata_rw_end(base, addr, drv_data->iordy, t0, t2, t2i); |
278 | return readl(base + IDEDATAIN); |
279 | } |
280 | |
281 | /* IDE register read */ |
282 | static u16 ep93xx_pata_read_reg(struct ep93xx_pata_data *drv_data, |
283 | unsigned long addr) |
284 | { |
285 | return ep93xx_pata_read(drv_data, addr, true); |
286 | } |
287 | |
288 | /* PIO data read */ |
289 | static u16 ep93xx_pata_read_data(struct ep93xx_pata_data *drv_data, |
290 | unsigned long addr) |
291 | { |
292 | return ep93xx_pata_read(drv_data, addr, false); |
293 | } |
294 | |
295 | static void ep93xx_pata_write(struct ep93xx_pata_data *drv_data, |
296 | u16 value, unsigned long addr, |
297 | bool reg) |
298 | { |
299 | void __iomem *base = drv_data->ide_base; |
300 | const struct ata_timing *t = &drv_data->t; |
301 | unsigned long t0 = reg ? t->cyc8b : t->cycle; |
302 | unsigned long t2 = reg ? t->act8b : t->active; |
303 | unsigned long t2i = reg ? t->rec8b : t->recover; |
304 | |
305 | ep93xx_pata_rw_begin(base, addr, t->setup); |
306 | /* |
307 | * Value from IDEDATAOUT register is driven onto the DD pins when |
308 | * DIOWN is low. (EP93xx UG p27-13) |
309 | */ |
310 | writel(value, base + IDEDATAOUT); |
311 | writel(IDECTRL_DIORN | addr, base + IDECTRL); |
312 | ep93xx_pata_rw_end(base, addr, drv_data->iordy, t0, t2, t2i); |
313 | } |
314 | |
315 | /* IDE register write */ |
316 | static void ep93xx_pata_write_reg(struct ep93xx_pata_data *drv_data, |
317 | u16 value, unsigned long addr) |
318 | { |
319 | ep93xx_pata_write(drv_data, value, addr, true); |
320 | } |
321 | |
322 | /* PIO data write */ |
323 | static void ep93xx_pata_write_data(struct ep93xx_pata_data *drv_data, |
324 | u16 value, unsigned long addr) |
325 | { |
326 | ep93xx_pata_write(drv_data, value, addr, false); |
327 | } |
328 | |
329 | static void ep93xx_pata_set_piomode(struct ata_port *ap, |
330 | struct ata_device *adev) |
331 | { |
332 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
333 | struct ata_device *pair = ata_dev_pair(adev); |
334 | /* |
335 | * Calculate timings for the delay loop, assuming ep93xx cpu speed |
336 | * is 200MHz (maximum possible for ep93xx). If actual cpu speed is |
337 | * slower, we will wait a bit longer in each delay. |
338 | * Additional division of cpu speed by 5, because single iteration |
339 | * of our delay loop takes 5 cpu cycles (25ns). |
340 | */ |
341 | unsigned long T = 1000000 / (200 / 5); |
342 | |
343 | ata_timing_compute(adev, adev->pio_mode, &drv_data->t, T, 0); |
344 | if (pair && pair->pio_mode) { |
345 | struct ata_timing t; |
346 | ata_timing_compute(pair, pair->pio_mode, &t, T, 0); |
347 | ata_timing_merge(&t, &drv_data->t, &drv_data->t, |
348 | ATA_TIMING_SETUP | ATA_TIMING_8BIT); |
349 | } |
350 | drv_data->iordy = ata_pio_need_iordy(adev); |
351 | |
352 | ep93xx_pata_enable_pio(drv_data->ide_base, |
353 | adev->pio_mode - XFER_PIO_0); |
354 | } |
355 | |
356 | /* Note: original code is ata_sff_check_status */ |
357 | static u8 ep93xx_pata_check_status(struct ata_port *ap) |
358 | { |
359 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
360 | |
361 | return ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_STATUS); |
362 | } |
363 | |
364 | static u8 ep93xx_pata_check_altstatus(struct ata_port *ap) |
365 | { |
366 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
367 | |
368 | return ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_ALTSTATUS); |
369 | } |
370 | |
371 | /* Note: original code is ata_sff_tf_load */ |
372 | static void ep93xx_pata_tf_load(struct ata_port *ap, |
373 | const struct ata_taskfile *tf) |
374 | { |
375 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
376 | unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; |
377 | |
378 | if (tf->ctl != ap->last_ctl) { |
379 | ep93xx_pata_write_reg(drv_data, tf->ctl, IDECTRL_ADDR_CTL); |
380 | ap->last_ctl = tf->ctl; |
381 | ata_wait_idle(ap); |
382 | } |
383 | |
384 | if (is_addr && (tf->flags & ATA_TFLAG_LBA48)) { |
385 | ep93xx_pata_write_reg(drv_data, tf->hob_feature, |
386 | IDECTRL_ADDR_FEATURE); |
387 | ep93xx_pata_write_reg(drv_data, tf->hob_nsect, |
388 | IDECTRL_ADDR_NSECT); |
389 | ep93xx_pata_write_reg(drv_data, tf->hob_lbal, |
390 | IDECTRL_ADDR_LBAL); |
391 | ep93xx_pata_write_reg(drv_data, tf->hob_lbam, |
392 | IDECTRL_ADDR_LBAM); |
393 | ep93xx_pata_write_reg(drv_data, tf->hob_lbah, |
394 | IDECTRL_ADDR_LBAH); |
395 | } |
396 | |
397 | if (is_addr) { |
398 | ep93xx_pata_write_reg(drv_data, tf->feature, |
399 | IDECTRL_ADDR_FEATURE); |
400 | ep93xx_pata_write_reg(drv_data, tf->nsect, IDECTRL_ADDR_NSECT); |
401 | ep93xx_pata_write_reg(drv_data, tf->lbal, IDECTRL_ADDR_LBAL); |
402 | ep93xx_pata_write_reg(drv_data, tf->lbam, IDECTRL_ADDR_LBAM); |
403 | ep93xx_pata_write_reg(drv_data, tf->lbah, IDECTRL_ADDR_LBAH); |
404 | } |
405 | |
406 | if (tf->flags & ATA_TFLAG_DEVICE) |
407 | ep93xx_pata_write_reg(drv_data, tf->device, |
408 | IDECTRL_ADDR_DEVICE); |
409 | |
410 | ata_wait_idle(ap); |
411 | } |
412 | |
413 | /* Note: original code is ata_sff_tf_read */ |
414 | static void ep93xx_pata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) |
415 | { |
416 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
417 | |
418 | tf->command = ep93xx_pata_check_status(ap); |
419 | tf->feature = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_FEATURE); |
420 | tf->nsect = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_NSECT); |
421 | tf->lbal = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAL); |
422 | tf->lbam = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAM); |
423 | tf->lbah = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAH); |
424 | tf->device = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_DEVICE); |
425 | |
426 | if (tf->flags & ATA_TFLAG_LBA48) { |
427 | ep93xx_pata_write_reg(drv_data, tf->ctl | ATA_HOB, |
428 | IDECTRL_ADDR_CTL); |
429 | tf->hob_feature = ep93xx_pata_read_reg(drv_data, |
430 | IDECTRL_ADDR_FEATURE); |
431 | tf->hob_nsect = ep93xx_pata_read_reg(drv_data, |
432 | IDECTRL_ADDR_NSECT); |
433 | tf->hob_lbal = ep93xx_pata_read_reg(drv_data, |
434 | IDECTRL_ADDR_LBAL); |
435 | tf->hob_lbam = ep93xx_pata_read_reg(drv_data, |
436 | IDECTRL_ADDR_LBAM); |
437 | tf->hob_lbah = ep93xx_pata_read_reg(drv_data, |
438 | IDECTRL_ADDR_LBAH); |
439 | ep93xx_pata_write_reg(drv_data, tf->ctl, IDECTRL_ADDR_CTL); |
440 | ap->last_ctl = tf->ctl; |
441 | } |
442 | } |
443 | |
444 | /* Note: original code is ata_sff_exec_command */ |
445 | static void ep93xx_pata_exec_command(struct ata_port *ap, |
446 | const struct ata_taskfile *tf) |
447 | { |
448 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
449 | |
450 | ep93xx_pata_write_reg(drv_data, tf->command, |
451 | IDECTRL_ADDR_COMMAND); |
452 | ata_sff_pause(ap); |
453 | } |
454 | |
455 | /* Note: original code is ata_sff_dev_select */ |
456 | static void ep93xx_pata_dev_select(struct ata_port *ap, unsigned int device) |
457 | { |
458 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
459 | u8 tmp = ATA_DEVICE_OBS; |
460 | |
461 | if (device != 0) |
462 | tmp |= ATA_DEV1; |
463 | |
464 | ep93xx_pata_write_reg(drv_data, tmp, IDECTRL_ADDR_DEVICE); |
465 | ata_sff_pause(ap); /* needed; also flushes, for mmio */ |
466 | } |
467 | |
468 | /* Note: original code is ata_sff_set_devctl */ |
469 | static void ep93xx_pata_set_devctl(struct ata_port *ap, u8 ctl) |
470 | { |
471 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
472 | |
473 | ep93xx_pata_write_reg(drv_data, ctl, IDECTRL_ADDR_CTL); |
474 | } |
475 | |
476 | /* Note: original code is ata_sff_data_xfer */ |
477 | static unsigned int ep93xx_pata_data_xfer(struct ata_device *adev, |
478 | unsigned char *buf, |
479 | unsigned int buflen, int rw) |
480 | { |
481 | struct ata_port *ap = adev->link->ap; |
482 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
483 | u16 *data = (u16 *)buf; |
484 | unsigned int words = buflen >> 1; |
485 | |
486 | /* Transfer multiple of 2 bytes */ |
487 | while (words--) |
488 | if (rw == READ) |
489 | *data++ = cpu_to_le16( |
490 | ep93xx_pata_read_data( |
491 | drv_data, IDECTRL_ADDR_DATA)); |
492 | else |
493 | ep93xx_pata_write_data(drv_data, le16_to_cpu(*data++), |
494 | IDECTRL_ADDR_DATA); |
495 | |
496 | /* Transfer trailing 1 byte, if any. */ |
497 | if (unlikely(buflen & 0x01)) { |
498 | unsigned char pad[2] = { }; |
499 | |
500 | buf += buflen - 1; |
501 | |
502 | if (rw == READ) { |
503 | *pad = cpu_to_le16( |
504 | ep93xx_pata_read_data( |
505 | drv_data, IDECTRL_ADDR_DATA)); |
506 | *buf = pad[0]; |
507 | } else { |
508 | pad[0] = *buf; |
509 | ep93xx_pata_write_data(drv_data, le16_to_cpu(*pad), |
510 | IDECTRL_ADDR_DATA); |
511 | } |
512 | words++; |
513 | } |
514 | |
515 | return words << 1; |
516 | } |
517 | |
518 | /* Note: original code is ata_devchk */ |
519 | static bool ep93xx_pata_device_is_present(struct ata_port *ap, |
520 | unsigned int device) |
521 | { |
522 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
523 | u8 nsect, lbal; |
524 | |
525 | ap->ops->sff_dev_select(ap, device); |
526 | |
527 | ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_NSECT); |
528 | ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_LBAL); |
529 | |
530 | ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_NSECT); |
531 | ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_LBAL); |
532 | |
533 | ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_NSECT); |
534 | ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_LBAL); |
535 | |
536 | nsect = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_NSECT); |
537 | lbal = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAL); |
538 | |
539 | if ((nsect == 0x55) && (lbal == 0xaa)) |
540 | return true; |
541 | |
542 | return false; |
543 | } |
544 | |
545 | /* Note: original code is ata_sff_wait_after_reset */ |
546 | static int ep93xx_pata_wait_after_reset(struct ata_link *link, |
547 | unsigned int devmask, |
548 | unsigned long deadline) |
549 | { |
550 | struct ata_port *ap = link->ap; |
551 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
552 | unsigned int dev0 = devmask & (1 << 0); |
553 | unsigned int dev1 = devmask & (1 << 1); |
554 | int rc, ret = 0; |
555 | |
556 | ata_msleep(ap, ATA_WAIT_AFTER_RESET); |
557 | |
558 | /* always check readiness of the master device */ |
559 | rc = ata_sff_wait_ready(link, deadline); |
560 | /* |
561 | * -ENODEV means the odd clown forgot the D7 pulldown resistor |
562 | * and TF status is 0xff, bail out on it too. |
563 | */ |
564 | if (rc) |
565 | return rc; |
566 | |
567 | /* |
568 | * if device 1 was found in ata_devchk, wait for register |
569 | * access briefly, then wait for BSY to clear. |
570 | */ |
571 | if (dev1) { |
572 | int i; |
573 | |
574 | ap->ops->sff_dev_select(ap, 1); |
575 | |
576 | /* |
577 | * Wait for register access. Some ATAPI devices fail |
578 | * to set nsect/lbal after reset, so don't waste too |
579 | * much time on it. We're gonna wait for !BSY anyway. |
580 | */ |
581 | for (i = 0; i < 2; i++) { |
582 | u8 nsect, lbal; |
583 | |
584 | nsect = ep93xx_pata_read_reg(drv_data, |
585 | IDECTRL_ADDR_NSECT); |
586 | lbal = ep93xx_pata_read_reg(drv_data, |
587 | IDECTRL_ADDR_LBAL); |
588 | if (nsect == 1 && lbal == 1) |
589 | break; |
590 | msleep(50); /* give drive a breather */ |
591 | } |
592 | |
593 | rc = ata_sff_wait_ready(link, deadline); |
594 | if (rc) { |
595 | if (rc != -ENODEV) |
596 | return rc; |
597 | ret = rc; |
598 | } |
599 | } |
600 | /* is all this really necessary? */ |
601 | ap->ops->sff_dev_select(ap, 0); |
602 | if (dev1) |
603 | ap->ops->sff_dev_select(ap, 1); |
604 | if (dev0) |
605 | ap->ops->sff_dev_select(ap, 0); |
606 | |
607 | return ret; |
608 | } |
609 | |
610 | /* Note: original code is ata_bus_softreset */ |
611 | static int ep93xx_pata_bus_softreset(struct ata_port *ap, unsigned int devmask, |
612 | unsigned long deadline) |
613 | { |
614 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
615 | |
616 | ep93xx_pata_write_reg(drv_data, ap->ctl, IDECTRL_ADDR_CTL); |
617 | udelay(20); /* FIXME: flush */ |
618 | ep93xx_pata_write_reg(drv_data, ap->ctl | ATA_SRST, IDECTRL_ADDR_CTL); |
619 | udelay(20); /* FIXME: flush */ |
620 | ep93xx_pata_write_reg(drv_data, ap->ctl, IDECTRL_ADDR_CTL); |
621 | ap->last_ctl = ap->ctl; |
622 | |
623 | return ep93xx_pata_wait_after_reset(&ap->link, devmask, deadline); |
624 | } |
625 | |
626 | static void ep93xx_pata_release_dma(struct ep93xx_pata_data *drv_data) |
627 | { |
628 | if (drv_data->dma_rx_channel) { |
629 | dma_release_channel(drv_data->dma_rx_channel); |
630 | drv_data->dma_rx_channel = NULL; |
631 | } |
632 | if (drv_data->dma_tx_channel) { |
633 | dma_release_channel(drv_data->dma_tx_channel); |
634 | drv_data->dma_tx_channel = NULL; |
635 | } |
636 | } |
637 | |
638 | static bool ep93xx_pata_dma_filter(struct dma_chan *chan, void *filter_param) |
639 | { |
640 | if (ep93xx_dma_chan_is_m2p(chan)) |
641 | return false; |
642 | |
643 | chan->private = filter_param; |
644 | return true; |
645 | } |
646 | |
647 | static void ep93xx_pata_dma_init(struct ep93xx_pata_data *drv_data) |
648 | { |
649 | const struct platform_device *pdev = drv_data->pdev; |
650 | dma_cap_mask_t mask; |
651 | struct dma_slave_config conf; |
652 | |
653 | dma_cap_zero(mask); |
654 | dma_cap_set(DMA_SLAVE, mask); |
655 | |
656 | /* |
657 | * Request two channels for IDE. Another possibility would be |
658 | * to request only one channel, and reprogram it's direction at |
659 | * start of new transfer. |
660 | */ |
661 | drv_data->dma_rx_data.port = EP93XX_DMA_IDE; |
662 | drv_data->dma_rx_data.direction = DMA_FROM_DEVICE; |
663 | drv_data->dma_rx_data.name = "ep93xx-pata-rx"; |
664 | drv_data->dma_rx_channel = dma_request_channel(mask, |
665 | ep93xx_pata_dma_filter, &drv_data->dma_rx_data); |
666 | if (!drv_data->dma_rx_channel) |
667 | return; |
668 | |
669 | drv_data->dma_tx_data.port = EP93XX_DMA_IDE; |
670 | drv_data->dma_tx_data.direction = DMA_TO_DEVICE; |
671 | drv_data->dma_tx_data.name = "ep93xx-pata-tx"; |
672 | drv_data->dma_tx_channel = dma_request_channel(mask, |
673 | ep93xx_pata_dma_filter, &drv_data->dma_tx_data); |
674 | if (!drv_data->dma_tx_channel) { |
675 | dma_release_channel(drv_data->dma_rx_channel); |
676 | return; |
677 | } |
678 | |
679 | /* Configure receive channel direction and source address */ |
680 | memset(&conf, 0, sizeof(conf)); |
681 | conf.direction = DMA_FROM_DEVICE; |
682 | conf.src_addr = drv_data->udma_in_phys; |
683 | conf.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; |
684 | if (dmaengine_slave_config(drv_data->dma_rx_channel, &conf)) { |
685 | dev_err(&pdev->dev, "failed to configure rx dma channel\n"); |
686 | ep93xx_pata_release_dma(drv_data); |
687 | return; |
688 | } |
689 | |
690 | /* Configure transmit channel direction and destination address */ |
691 | memset(&conf, 0, sizeof(conf)); |
692 | conf.direction = DMA_TO_DEVICE; |
693 | conf.dst_addr = drv_data->udma_out_phys; |
694 | conf.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; |
695 | if (dmaengine_slave_config(drv_data->dma_tx_channel, &conf)) { |
696 | dev_err(&pdev->dev, "failed to configure tx dma channel\n"); |
697 | ep93xx_pata_release_dma(drv_data); |
698 | } |
699 | } |
700 | |
701 | static void ep93xx_pata_dma_start(struct ata_queued_cmd *qc) |
702 | { |
703 | struct dma_async_tx_descriptor *txd; |
704 | struct ep93xx_pata_data *drv_data = qc->ap->host->private_data; |
705 | void __iomem *base = drv_data->ide_base; |
706 | struct ata_device *adev = qc->dev; |
707 | u32 v = qc->dma_dir == DMA_TO_DEVICE ? IDEUDMAOP_RWOP : 0; |
708 | struct dma_chan *channel = qc->dma_dir == DMA_TO_DEVICE |
709 | ? drv_data->dma_tx_channel : drv_data->dma_rx_channel; |
710 | |
711 | txd = channel->device->device_prep_slave_sg(channel, qc->sg, |
712 | qc->n_elem, qc->dma_dir, DMA_CTRL_ACK, NULL); |
713 | if (!txd) { |
714 | dev_err(qc->ap->dev, "failed to prepare slave for sg dma\n"); |
715 | return; |
716 | } |
717 | txd->callback = NULL; |
718 | txd->callback_param = NULL; |
719 | |
720 | if (dmaengine_submit(txd) < 0) { |
721 | dev_err(qc->ap->dev, "failed to submit dma transfer\n"); |
722 | return; |
723 | } |
724 | dma_async_issue_pending(channel); |
725 | |
726 | /* |
727 | * When enabling UDMA operation, IDEUDMAOP register needs to be |
728 | * programmed in three step sequence: |
729 | * 1) set or clear the RWOP bit, |
730 | * 2) perform dummy read of the register, |
731 | * 3) set the UEN bit. |
732 | */ |
733 | writel(v, base + IDEUDMAOP); |
734 | readl(base + IDEUDMAOP); |
735 | writel(v | IDEUDMAOP_UEN, base + IDEUDMAOP); |
736 | |
737 | writel(IDECFG_IDEEN | IDECFG_UDMA | |
738 | ((adev->xfer_mode - XFER_UDMA_0) << IDECFG_MODE_SHIFT), |
739 | base + IDECFG); |
740 | } |
741 | |
742 | static void ep93xx_pata_dma_stop(struct ata_queued_cmd *qc) |
743 | { |
744 | struct ep93xx_pata_data *drv_data = qc->ap->host->private_data; |
745 | void __iomem *base = drv_data->ide_base; |
746 | |
747 | /* terminate all dma transfers, if not yet finished */ |
748 | dmaengine_terminate_all(drv_data->dma_rx_channel); |
749 | dmaengine_terminate_all(drv_data->dma_tx_channel); |
750 | |
751 | /* |
752 | * To properly stop IDE-DMA, IDEUDMAOP register must to be cleared |
753 | * and IDECTRL register must be set to default value. |
754 | */ |
755 | writel(0, base + IDEUDMAOP); |
756 | writel(readl(base + IDECTRL) | IDECTRL_DIOWN | IDECTRL_DIORN | |
757 | IDECTRL_CS0N | IDECTRL_CS1N, base + IDECTRL); |
758 | |
759 | ep93xx_pata_enable_pio(drv_data->ide_base, |
760 | qc->dev->pio_mode - XFER_PIO_0); |
761 | |
762 | ata_sff_dma_pause(qc->ap); |
763 | } |
764 | |
765 | static void ep93xx_pata_dma_setup(struct ata_queued_cmd *qc) |
766 | { |
767 | qc->ap->ops->sff_exec_command(qc->ap, &qc->tf); |
768 | } |
769 | |
770 | static u8 ep93xx_pata_dma_status(struct ata_port *ap) |
771 | { |
772 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
773 | u32 val = readl(drv_data->ide_base + IDEUDMASTS); |
774 | |
775 | /* |
776 | * UDMA Status Register bits: |
777 | * |
778 | * DMAIDE - DMA request signal from UDMA state machine, |
779 | * INTIDE - INT line generated by UDMA because of errors in the |
780 | * state machine, |
781 | * SBUSY - UDMA state machine busy, not in idle state, |
782 | * NDO - error for data-out not completed, |
783 | * NDI - error for data-in not completed, |
784 | * N4X - error for data transferred not multiplies of four |
785 | * 32-bit words. |
786 | * (EP93xx UG p27-17) |
787 | */ |
788 | if (val & IDEUDMASTS_NDO || val & IDEUDMASTS_NDI || |
789 | val & IDEUDMASTS_N4X || val & IDEUDMASTS_INTIDE) |
790 | return ATA_DMA_ERR; |
791 | |
792 | /* read INTRQ (INT[3]) pin input state */ |
793 | if (readl(drv_data->ide_base + IDECTRL) & IDECTRL_INTRQ) |
794 | return ATA_DMA_INTR; |
795 | |
796 | if (val & IDEUDMASTS_SBUSY || val & IDEUDMASTS_DMAIDE) |
797 | return ATA_DMA_ACTIVE; |
798 | |
799 | return 0; |
800 | } |
801 | |
802 | /* Note: original code is ata_sff_softreset */ |
803 | static int ep93xx_pata_softreset(struct ata_link *al, unsigned int *classes, |
804 | unsigned long deadline) |
805 | { |
806 | struct ata_port *ap = al->ap; |
807 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; |
808 | unsigned int devmask = 0; |
809 | int rc; |
810 | u8 err; |
811 | |
812 | /* determine if device 0/1 are present */ |
813 | if (ep93xx_pata_device_is_present(ap, 0)) |
814 | devmask |= (1 << 0); |
815 | if (slave_possible && ep93xx_pata_device_is_present(ap, 1)) |
816 | devmask |= (1 << 1); |
817 | |
818 | /* select device 0 again */ |
819 | ap->ops->sff_dev_select(al->ap, 0); |
820 | |
821 | /* issue bus reset */ |
822 | rc = ep93xx_pata_bus_softreset(ap, devmask, deadline); |
823 | /* if link is ocuppied, -ENODEV too is an error */ |
824 | if (rc && (rc != -ENODEV || sata_scr_valid(al))) { |
825 | ata_link_printk(al, KERN_ERR, "SRST failed (errno=%d)\n", |
826 | rc); |
827 | return rc; |
828 | } |
829 | |
830 | /* determine by signature whether we have ATA or ATAPI devices */ |
831 | classes[0] = ata_sff_dev_classify(&al->device[0], devmask & (1 << 0), |
832 | &err); |
833 | if (slave_possible && err != 0x81) |
834 | classes[1] = ata_sff_dev_classify(&al->device[1], |
835 | devmask & (1 << 1), &err); |
836 | |
837 | return 0; |
838 | } |
839 | |
840 | /* Note: original code is ata_sff_drain_fifo */ |
841 | static void ep93xx_pata_drain_fifo(struct ata_queued_cmd *qc) |
842 | { |
843 | int count; |
844 | struct ata_port *ap; |
845 | struct ep93xx_pata_data *drv_data; |
846 | |
847 | /* We only need to flush incoming data when a command was running */ |
848 | if (qc == NULL || qc->dma_dir == DMA_TO_DEVICE) |
849 | return; |
850 | |
851 | ap = qc->ap; |
852 | drv_data = ap->host->private_data; |
853 | /* Drain up to 64K of data before we give up this recovery method */ |
854 | for (count = 0; (ap->ops->sff_check_status(ap) & ATA_DRQ) |
855 | && count < 65536; count += 2) |
856 | ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_DATA); |
857 | |
858 | /* Can become DEBUG later */ |
859 | if (count) |
860 | ata_port_printk(ap, KERN_DEBUG, |
861 | "drained %d bytes to clear DRQ.\n", count); |
862 | |
863 | } |
864 | |
865 | static int ep93xx_pata_port_start(struct ata_port *ap) |
866 | { |
867 | struct ep93xx_pata_data *drv_data = ap->host->private_data; |
868 | |
869 | /* |
870 | * Set timings to safe values at startup (= number of ns from ATA |
871 | * specification), we'll switch to properly calculated values later. |
872 | */ |
873 | drv_data->t = *ata_timing_find_mode(XFER_PIO_0); |
874 | return 0; |
875 | } |
876 | |
877 | static struct scsi_host_template ep93xx_pata_sht = { |
878 | ATA_BASE_SHT(DRV_NAME), |
879 | /* ep93xx dma implementation limit */ |
880 | .sg_tablesize = 32, |
881 | /* ep93xx dma can't transfer 65536 bytes at once */ |
882 | .dma_boundary = 0x7fff, |
883 | }; |
884 | |
885 | static struct ata_port_operations ep93xx_pata_port_ops = { |
886 | .inherits = &ata_bmdma_port_ops, |
887 | |
888 | .qc_prep = ata_noop_qc_prep, |
889 | |
890 | .softreset = ep93xx_pata_softreset, |
891 | .hardreset = ATA_OP_NULL, |
892 | |
893 | .sff_dev_select = ep93xx_pata_dev_select, |
894 | .sff_set_devctl = ep93xx_pata_set_devctl, |
895 | .sff_check_status = ep93xx_pata_check_status, |
896 | .sff_check_altstatus = ep93xx_pata_check_altstatus, |
897 | .sff_tf_load = ep93xx_pata_tf_load, |
898 | .sff_tf_read = ep93xx_pata_tf_read, |
899 | .sff_exec_command = ep93xx_pata_exec_command, |
900 | .sff_data_xfer = ep93xx_pata_data_xfer, |
901 | .sff_drain_fifo = ep93xx_pata_drain_fifo, |
902 | .sff_irq_clear = ATA_OP_NULL, |
903 | |
904 | .set_piomode = ep93xx_pata_set_piomode, |
905 | |
906 | .bmdma_setup = ep93xx_pata_dma_setup, |
907 | .bmdma_start = ep93xx_pata_dma_start, |
908 | .bmdma_stop = ep93xx_pata_dma_stop, |
909 | .bmdma_status = ep93xx_pata_dma_status, |
910 | |
911 | .cable_detect = ata_cable_unknown, |
912 | .port_start = ep93xx_pata_port_start, |
913 | }; |
914 | |
915 | static int __devinit ep93xx_pata_probe(struct platform_device *pdev) |
916 | { |
917 | struct ep93xx_pata_data *drv_data; |
918 | struct ata_host *host; |
919 | struct ata_port *ap; |
920 | unsigned int irq; |
921 | struct resource *mem_res; |
922 | void __iomem *ide_base; |
923 | int err; |
924 | |
925 | err = ep93xx_ide_acquire_gpio(pdev); |
926 | if (err) |
927 | return err; |
928 | |
929 | /* INT[3] (IRQ_EP93XX_EXT3) line connected as pull down */ |
930 | irq = platform_get_irq(pdev, 0); |
931 | if (irq < 0) { |
932 | err = -ENXIO; |
933 | goto err_rel_gpio; |
934 | } |
935 | |
936 | mem_res = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
937 | if (!mem_res) { |
938 | err = -ENXIO; |
939 | goto err_rel_gpio; |
940 | } |
941 | |
942 | ide_base = devm_request_and_ioremap(&pdev->dev, mem_res); |
943 | if (!ide_base) { |
944 | err = -ENXIO; |
945 | goto err_rel_gpio; |
946 | } |
947 | |
948 | drv_data = devm_kzalloc(&pdev->dev, sizeof(*drv_data), GFP_KERNEL); |
949 | if (!drv_data) { |
950 | err = -ENXIO; |
951 | goto err_rel_gpio; |
952 | } |
953 | |
954 | platform_set_drvdata(pdev, drv_data); |
955 | drv_data->pdev = pdev; |
956 | drv_data->ide_base = ide_base; |
957 | drv_data->udma_in_phys = mem_res->start + IDEUDMADATAIN; |
958 | drv_data->udma_out_phys = mem_res->start + IDEUDMADATAOUT; |
959 | ep93xx_pata_dma_init(drv_data); |
960 | |
961 | /* allocate host */ |
962 | host = ata_host_alloc(&pdev->dev, 1); |
963 | if (!host) { |
964 | err = -ENXIO; |
965 | goto err_rel_dma; |
966 | } |
967 | |
968 | ep93xx_pata_clear_regs(ide_base); |
969 | |
970 | host->private_data = drv_data; |
971 | |
972 | ap = host->ports[0]; |
973 | ap->dev = &pdev->dev; |
974 | ap->ops = &ep93xx_pata_port_ops; |
975 | ap->flags |= ATA_FLAG_SLAVE_POSS; |
976 | ap->pio_mask = ATA_PIO4; |
977 | |
978 | /* |
979 | * Maximum UDMA modes: |
980 | * EP931x rev.E0 - UDMA2 |
981 | * EP931x rev.E1 - UDMA3 |
982 | * EP931x rev.E2 - UDMA4 |
983 | * |
984 | * MWDMA support was removed from EP931x rev.E2, |
985 | * so this driver supports only UDMA modes. |
986 | */ |
987 | if (drv_data->dma_rx_channel && drv_data->dma_tx_channel) { |
988 | int chip_rev = ep93xx_chip_revision(); |
989 | |
990 | if (chip_rev == EP93XX_CHIP_REV_E1) |
991 | ap->udma_mask = ATA_UDMA3; |
992 | else if (chip_rev == EP93XX_CHIP_REV_E2) |
993 | ap->udma_mask = ATA_UDMA4; |
994 | else |
995 | ap->udma_mask = ATA_UDMA2; |
996 | } |
997 | |
998 | /* defaults, pio 0 */ |
999 | ep93xx_pata_enable_pio(ide_base, 0); |
1000 | |
1001 | dev_info(&pdev->dev, "version " DRV_VERSION "\n"); |
1002 | |
1003 | /* activate host */ |
1004 | err = ata_host_activate(host, irq, ata_bmdma_interrupt, 0, |
1005 | &ep93xx_pata_sht); |
1006 | if (err == 0) |
1007 | return 0; |
1008 | |
1009 | err_rel_dma: |
1010 | ep93xx_pata_release_dma(drv_data); |
1011 | err_rel_gpio: |
1012 | ep93xx_ide_release_gpio(pdev); |
1013 | return err; |
1014 | } |
1015 | |
1016 | static int __devexit ep93xx_pata_remove(struct platform_device *pdev) |
1017 | { |
1018 | struct ata_host *host = platform_get_drvdata(pdev); |
1019 | struct ep93xx_pata_data *drv_data = host->private_data; |
1020 | |
1021 | ata_host_detach(host); |
1022 | ep93xx_pata_release_dma(drv_data); |
1023 | ep93xx_pata_clear_regs(drv_data->ide_base); |
1024 | ep93xx_ide_release_gpio(pdev); |
1025 | return 0; |
1026 | } |
1027 | |
1028 | static struct platform_driver ep93xx_pata_platform_driver = { |
1029 | .driver = { |
1030 | .name = DRV_NAME, |
1031 | .owner = THIS_MODULE, |
1032 | }, |
1033 | .probe = ep93xx_pata_probe, |
1034 | .remove = __devexit_p(ep93xx_pata_remove), |
1035 | }; |
1036 | |
1037 | module_platform_driver(ep93xx_pata_platform_driver); |
1038 | |
1039 | MODULE_AUTHOR("Alessandro Zummo, Lennert Buytenhek, Joao Ramos, " |
1040 | "Bartlomiej Zolnierkiewicz, Rafal Prylowski"); |
1041 | MODULE_DESCRIPTION("low-level driver for cirrus ep93xx IDE controller"); |
1042 | MODULE_LICENSE("GPL"); |
1043 | MODULE_VERSION(DRV_VERSION); |
1044 | MODULE_ALIAS("platform:pata_ep93xx"); |
1045 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9