Create platform_device.h to contain all the platform device details.
[deliverable/linux.git] / drivers / mmc / wbsd.c
1 /*
2 * linux/drivers/mmc/wbsd.c - Winbond W83L51xD SD/MMC driver
3 *
4 * Copyright (C) 2004-2005 Pierre Ossman, All Rights Reserved.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 as
8 * published by the Free Software Foundation.
9 *
10 *
11 * Warning!
12 *
13 * Changes to the FIFO system should be done with extreme care since
14 * the hardware is full of bugs related to the FIFO. Known issues are:
15 *
16 * - FIFO size field in FSR is always zero.
17 *
18 * - FIFO interrupts tend not to work as they should. Interrupts are
19 * triggered only for full/empty events, not for threshold values.
20 *
21 * - On APIC systems the FIFO empty interrupt is sometimes lost.
22 */
23
24 #include <linux/config.h>
25 #include <linux/module.h>
26 #include <linux/moduleparam.h>
27 #include <linux/init.h>
28 #include <linux/ioport.h>
29 #include <linux/platform_device.h>
30 #include <linux/interrupt.h>
31 #include <linux/dma-mapping.h>
32 #include <linux/delay.h>
33 #include <linux/pnp.h>
34 #include <linux/highmem.h>
35 #include <linux/mmc/host.h>
36 #include <linux/mmc/protocol.h>
37
38 #include <asm/io.h>
39 #include <asm/dma.h>
40 #include <asm/scatterlist.h>
41
42 #include "wbsd.h"
43
44 #define DRIVER_NAME "wbsd"
45 #define DRIVER_VERSION "1.4"
46
47 #ifdef CONFIG_MMC_DEBUG
48 #define DBG(x...) \
49 printk(KERN_DEBUG DRIVER_NAME ": " x)
50 #define DBGF(f, x...) \
51 printk(KERN_DEBUG DRIVER_NAME " [%s()]: " f, __func__ , ##x)
52 #else
53 #define DBG(x...) do { } while (0)
54 #define DBGF(x...) do { } while (0)
55 #endif
56
57 /*
58 * Device resources
59 */
60
61 #ifdef CONFIG_PNP
62
63 static const struct pnp_device_id pnp_dev_table[] = {
64 { "WEC0517", 0 },
65 { "WEC0518", 0 },
66 { "", 0 },
67 };
68
69 MODULE_DEVICE_TABLE(pnp, pnp_dev_table);
70
71 #endif /* CONFIG_PNP */
72
73 static const int config_ports[] = { 0x2E, 0x4E };
74 static const int unlock_codes[] = { 0x83, 0x87 };
75
76 static const int valid_ids[] = {
77 0x7112,
78 };
79
80 #ifdef CONFIG_PNP
81 static unsigned int nopnp = 0;
82 #else
83 static const unsigned int nopnp = 1;
84 #endif
85 static unsigned int io = 0x248;
86 static unsigned int irq = 6;
87 static int dma = 2;
88
89 /*
90 * Basic functions
91 */
92
93 static inline void wbsd_unlock_config(struct wbsd_host* host)
94 {
95 BUG_ON(host->config == 0);
96
97 outb(host->unlock_code, host->config);
98 outb(host->unlock_code, host->config);
99 }
100
101 static inline void wbsd_lock_config(struct wbsd_host* host)
102 {
103 BUG_ON(host->config == 0);
104
105 outb(LOCK_CODE, host->config);
106 }
107
108 static inline void wbsd_write_config(struct wbsd_host* host, u8 reg, u8 value)
109 {
110 BUG_ON(host->config == 0);
111
112 outb(reg, host->config);
113 outb(value, host->config + 1);
114 }
115
116 static inline u8 wbsd_read_config(struct wbsd_host* host, u8 reg)
117 {
118 BUG_ON(host->config == 0);
119
120 outb(reg, host->config);
121 return inb(host->config + 1);
122 }
123
124 static inline void wbsd_write_index(struct wbsd_host* host, u8 index, u8 value)
125 {
126 outb(index, host->base + WBSD_IDXR);
127 outb(value, host->base + WBSD_DATAR);
128 }
129
130 static inline u8 wbsd_read_index(struct wbsd_host* host, u8 index)
131 {
132 outb(index, host->base + WBSD_IDXR);
133 return inb(host->base + WBSD_DATAR);
134 }
135
136 /*
137 * Common routines
138 */
139
140 static void wbsd_init_device(struct wbsd_host* host)
141 {
142 u8 setup, ier;
143
144 /*
145 * Reset chip (SD/MMC part) and fifo.
146 */
147 setup = wbsd_read_index(host, WBSD_IDX_SETUP);
148 setup |= WBSD_FIFO_RESET | WBSD_SOFT_RESET;
149 wbsd_write_index(host, WBSD_IDX_SETUP, setup);
150
151 /*
152 * Set DAT3 to input
153 */
154 setup &= ~WBSD_DAT3_H;
155 wbsd_write_index(host, WBSD_IDX_SETUP, setup);
156 host->flags &= ~WBSD_FIGNORE_DETECT;
157
158 /*
159 * Read back default clock.
160 */
161 host->clk = wbsd_read_index(host, WBSD_IDX_CLK);
162
163 /*
164 * Power down port.
165 */
166 outb(WBSD_POWER_N, host->base + WBSD_CSR);
167
168 /*
169 * Set maximum timeout.
170 */
171 wbsd_write_index(host, WBSD_IDX_TAAC, 0x7F);
172
173 /*
174 * Test for card presence
175 */
176 if (inb(host->base + WBSD_CSR) & WBSD_CARDPRESENT)
177 host->flags |= WBSD_FCARD_PRESENT;
178 else
179 host->flags &= ~WBSD_FCARD_PRESENT;
180
181 /*
182 * Enable interesting interrupts.
183 */
184 ier = 0;
185 ier |= WBSD_EINT_CARD;
186 ier |= WBSD_EINT_FIFO_THRE;
187 ier |= WBSD_EINT_CCRC;
188 ier |= WBSD_EINT_TIMEOUT;
189 ier |= WBSD_EINT_CRC;
190 ier |= WBSD_EINT_TC;
191
192 outb(ier, host->base + WBSD_EIR);
193
194 /*
195 * Clear interrupts.
196 */
197 inb(host->base + WBSD_ISR);
198 }
199
200 static void wbsd_reset(struct wbsd_host* host)
201 {
202 u8 setup;
203
204 printk(KERN_ERR DRIVER_NAME ": Resetting chip\n");
205
206 /*
207 * Soft reset of chip (SD/MMC part).
208 */
209 setup = wbsd_read_index(host, WBSD_IDX_SETUP);
210 setup |= WBSD_SOFT_RESET;
211 wbsd_write_index(host, WBSD_IDX_SETUP, setup);
212 }
213
214 static void wbsd_request_end(struct wbsd_host* host, struct mmc_request* mrq)
215 {
216 unsigned long dmaflags;
217
218 DBGF("Ending request, cmd (%x)\n", mrq->cmd->opcode);
219
220 if (host->dma >= 0)
221 {
222 /*
223 * Release ISA DMA controller.
224 */
225 dmaflags = claim_dma_lock();
226 disable_dma(host->dma);
227 clear_dma_ff(host->dma);
228 release_dma_lock(dmaflags);
229
230 /*
231 * Disable DMA on host.
232 */
233 wbsd_write_index(host, WBSD_IDX_DMA, 0);
234 }
235
236 host->mrq = NULL;
237
238 /*
239 * MMC layer might call back into the driver so first unlock.
240 */
241 spin_unlock(&host->lock);
242 mmc_request_done(host->mmc, mrq);
243 spin_lock(&host->lock);
244 }
245
246 /*
247 * Scatter/gather functions
248 */
249
250 static inline void wbsd_init_sg(struct wbsd_host* host, struct mmc_data* data)
251 {
252 /*
253 * Get info. about SG list from data structure.
254 */
255 host->cur_sg = data->sg;
256 host->num_sg = data->sg_len;
257
258 host->offset = 0;
259 host->remain = host->cur_sg->length;
260 }
261
262 static inline int wbsd_next_sg(struct wbsd_host* host)
263 {
264 /*
265 * Skip to next SG entry.
266 */
267 host->cur_sg++;
268 host->num_sg--;
269
270 /*
271 * Any entries left?
272 */
273 if (host->num_sg > 0)
274 {
275 host->offset = 0;
276 host->remain = host->cur_sg->length;
277 }
278
279 return host->num_sg;
280 }
281
282 static inline char* wbsd_kmap_sg(struct wbsd_host* host)
283 {
284 host->mapped_sg = kmap_atomic(host->cur_sg->page, KM_BIO_SRC_IRQ) +
285 host->cur_sg->offset;
286 return host->mapped_sg;
287 }
288
289 static inline void wbsd_kunmap_sg(struct wbsd_host* host)
290 {
291 kunmap_atomic(host->mapped_sg, KM_BIO_SRC_IRQ);
292 }
293
294 static inline void wbsd_sg_to_dma(struct wbsd_host* host, struct mmc_data* data)
295 {
296 unsigned int len, i, size;
297 struct scatterlist* sg;
298 char* dmabuf = host->dma_buffer;
299 char* sgbuf;
300
301 size = host->size;
302
303 sg = data->sg;
304 len = data->sg_len;
305
306 /*
307 * Just loop through all entries. Size might not
308 * be the entire list though so make sure that
309 * we do not transfer too much.
310 */
311 for (i = 0;i < len;i++)
312 {
313 sgbuf = kmap_atomic(sg[i].page, KM_BIO_SRC_IRQ) + sg[i].offset;
314 if (size < sg[i].length)
315 memcpy(dmabuf, sgbuf, size);
316 else
317 memcpy(dmabuf, sgbuf, sg[i].length);
318 kunmap_atomic(sgbuf, KM_BIO_SRC_IRQ);
319 dmabuf += sg[i].length;
320
321 if (size < sg[i].length)
322 size = 0;
323 else
324 size -= sg[i].length;
325
326 if (size == 0)
327 break;
328 }
329
330 /*
331 * Check that we didn't get a request to transfer
332 * more data than can fit into the SG list.
333 */
334
335 BUG_ON(size != 0);
336
337 host->size -= size;
338 }
339
340 static inline void wbsd_dma_to_sg(struct wbsd_host* host, struct mmc_data* data)
341 {
342 unsigned int len, i, size;
343 struct scatterlist* sg;
344 char* dmabuf = host->dma_buffer;
345 char* sgbuf;
346
347 size = host->size;
348
349 sg = data->sg;
350 len = data->sg_len;
351
352 /*
353 * Just loop through all entries. Size might not
354 * be the entire list though so make sure that
355 * we do not transfer too much.
356 */
357 for (i = 0;i < len;i++)
358 {
359 sgbuf = kmap_atomic(sg[i].page, KM_BIO_SRC_IRQ) + sg[i].offset;
360 if (size < sg[i].length)
361 memcpy(sgbuf, dmabuf, size);
362 else
363 memcpy(sgbuf, dmabuf, sg[i].length);
364 kunmap_atomic(sgbuf, KM_BIO_SRC_IRQ);
365 dmabuf += sg[i].length;
366
367 if (size < sg[i].length)
368 size = 0;
369 else
370 size -= sg[i].length;
371
372 if (size == 0)
373 break;
374 }
375
376 /*
377 * Check that we didn't get a request to transfer
378 * more data than can fit into the SG list.
379 */
380
381 BUG_ON(size != 0);
382
383 host->size -= size;
384 }
385
386 /*
387 * Command handling
388 */
389
390 static inline void wbsd_get_short_reply(struct wbsd_host* host,
391 struct mmc_command* cmd)
392 {
393 /*
394 * Correct response type?
395 */
396 if (wbsd_read_index(host, WBSD_IDX_RSPLEN) != WBSD_RSP_SHORT)
397 {
398 cmd->error = MMC_ERR_INVALID;
399 return;
400 }
401
402 cmd->resp[0] =
403 wbsd_read_index(host, WBSD_IDX_RESP12) << 24;
404 cmd->resp[0] |=
405 wbsd_read_index(host, WBSD_IDX_RESP13) << 16;
406 cmd->resp[0] |=
407 wbsd_read_index(host, WBSD_IDX_RESP14) << 8;
408 cmd->resp[0] |=
409 wbsd_read_index(host, WBSD_IDX_RESP15) << 0;
410 cmd->resp[1] =
411 wbsd_read_index(host, WBSD_IDX_RESP16) << 24;
412 }
413
414 static inline void wbsd_get_long_reply(struct wbsd_host* host,
415 struct mmc_command* cmd)
416 {
417 int i;
418
419 /*
420 * Correct response type?
421 */
422 if (wbsd_read_index(host, WBSD_IDX_RSPLEN) != WBSD_RSP_LONG)
423 {
424 cmd->error = MMC_ERR_INVALID;
425 return;
426 }
427
428 for (i = 0;i < 4;i++)
429 {
430 cmd->resp[i] =
431 wbsd_read_index(host, WBSD_IDX_RESP1 + i * 4) << 24;
432 cmd->resp[i] |=
433 wbsd_read_index(host, WBSD_IDX_RESP2 + i * 4) << 16;
434 cmd->resp[i] |=
435 wbsd_read_index(host, WBSD_IDX_RESP3 + i * 4) << 8;
436 cmd->resp[i] |=
437 wbsd_read_index(host, WBSD_IDX_RESP4 + i * 4) << 0;
438 }
439 }
440
441 static void wbsd_send_command(struct wbsd_host* host, struct mmc_command* cmd)
442 {
443 int i;
444 u8 status, isr;
445
446 DBGF("Sending cmd (%x)\n", cmd->opcode);
447
448 /*
449 * Clear accumulated ISR. The interrupt routine
450 * will fill this one with events that occur during
451 * transfer.
452 */
453 host->isr = 0;
454
455 /*
456 * Send the command (CRC calculated by host).
457 */
458 outb(cmd->opcode, host->base + WBSD_CMDR);
459 for (i = 3;i >= 0;i--)
460 outb((cmd->arg >> (i * 8)) & 0xff, host->base + WBSD_CMDR);
461
462 cmd->error = MMC_ERR_NONE;
463
464 /*
465 * Wait for the request to complete.
466 */
467 do {
468 status = wbsd_read_index(host, WBSD_IDX_STATUS);
469 } while (status & WBSD_CARDTRAFFIC);
470
471 /*
472 * Do we expect a reply?
473 */
474 if ((cmd->flags & MMC_RSP_MASK) != MMC_RSP_NONE)
475 {
476 /*
477 * Read back status.
478 */
479 isr = host->isr;
480
481 /* Card removed? */
482 if (isr & WBSD_INT_CARD)
483 cmd->error = MMC_ERR_TIMEOUT;
484 /* Timeout? */
485 else if (isr & WBSD_INT_TIMEOUT)
486 cmd->error = MMC_ERR_TIMEOUT;
487 /* CRC? */
488 else if ((cmd->flags & MMC_RSP_CRC) && (isr & WBSD_INT_CRC))
489 cmd->error = MMC_ERR_BADCRC;
490 /* All ok */
491 else
492 {
493 if ((cmd->flags & MMC_RSP_MASK) == MMC_RSP_SHORT)
494 wbsd_get_short_reply(host, cmd);
495 else
496 wbsd_get_long_reply(host, cmd);
497 }
498 }
499
500 DBGF("Sent cmd (%x), res %d\n", cmd->opcode, cmd->error);
501 }
502
503 /*
504 * Data functions
505 */
506
507 static void wbsd_empty_fifo(struct wbsd_host* host)
508 {
509 struct mmc_data* data = host->mrq->cmd->data;
510 char* buffer;
511 int i, fsr, fifo;
512
513 /*
514 * Handle excessive data.
515 */
516 if (data->bytes_xfered == host->size)
517 return;
518
519 buffer = wbsd_kmap_sg(host) + host->offset;
520
521 /*
522 * Drain the fifo. This has a tendency to loop longer
523 * than the FIFO length (usually one block).
524 */
525 while (!((fsr = inb(host->base + WBSD_FSR)) & WBSD_FIFO_EMPTY))
526 {
527 /*
528 * The size field in the FSR is broken so we have to
529 * do some guessing.
530 */
531 if (fsr & WBSD_FIFO_FULL)
532 fifo = 16;
533 else if (fsr & WBSD_FIFO_FUTHRE)
534 fifo = 8;
535 else
536 fifo = 1;
537
538 for (i = 0;i < fifo;i++)
539 {
540 *buffer = inb(host->base + WBSD_DFR);
541 buffer++;
542 host->offset++;
543 host->remain--;
544
545 data->bytes_xfered++;
546
547 /*
548 * Transfer done?
549 */
550 if (data->bytes_xfered == host->size)
551 {
552 wbsd_kunmap_sg(host);
553 return;
554 }
555
556 /*
557 * End of scatter list entry?
558 */
559 if (host->remain == 0)
560 {
561 wbsd_kunmap_sg(host);
562
563 /*
564 * Get next entry. Check if last.
565 */
566 if (!wbsd_next_sg(host))
567 {
568 /*
569 * We should never reach this point.
570 * It means that we're trying to
571 * transfer more blocks than can fit
572 * into the scatter list.
573 */
574 BUG_ON(1);
575
576 host->size = data->bytes_xfered;
577
578 return;
579 }
580
581 buffer = wbsd_kmap_sg(host);
582 }
583 }
584 }
585
586 wbsd_kunmap_sg(host);
587
588 /*
589 * This is a very dirty hack to solve a
590 * hardware problem. The chip doesn't trigger
591 * FIFO threshold interrupts properly.
592 */
593 if ((host->size - data->bytes_xfered) < 16)
594 tasklet_schedule(&host->fifo_tasklet);
595 }
596
597 static void wbsd_fill_fifo(struct wbsd_host* host)
598 {
599 struct mmc_data* data = host->mrq->cmd->data;
600 char* buffer;
601 int i, fsr, fifo;
602
603 /*
604 * Check that we aren't being called after the
605 * entire buffer has been transfered.
606 */
607 if (data->bytes_xfered == host->size)
608 return;
609
610 buffer = wbsd_kmap_sg(host) + host->offset;
611
612 /*
613 * Fill the fifo. This has a tendency to loop longer
614 * than the FIFO length (usually one block).
615 */
616 while (!((fsr = inb(host->base + WBSD_FSR)) & WBSD_FIFO_FULL))
617 {
618 /*
619 * The size field in the FSR is broken so we have to
620 * do some guessing.
621 */
622 if (fsr & WBSD_FIFO_EMPTY)
623 fifo = 0;
624 else if (fsr & WBSD_FIFO_EMTHRE)
625 fifo = 8;
626 else
627 fifo = 15;
628
629 for (i = 16;i > fifo;i--)
630 {
631 outb(*buffer, host->base + WBSD_DFR);
632 buffer++;
633 host->offset++;
634 host->remain--;
635
636 data->bytes_xfered++;
637
638 /*
639 * Transfer done?
640 */
641 if (data->bytes_xfered == host->size)
642 {
643 wbsd_kunmap_sg(host);
644 return;
645 }
646
647 /*
648 * End of scatter list entry?
649 */
650 if (host->remain == 0)
651 {
652 wbsd_kunmap_sg(host);
653
654 /*
655 * Get next entry. Check if last.
656 */
657 if (!wbsd_next_sg(host))
658 {
659 /*
660 * We should never reach this point.
661 * It means that we're trying to
662 * transfer more blocks than can fit
663 * into the scatter list.
664 */
665 BUG_ON(1);
666
667 host->size = data->bytes_xfered;
668
669 return;
670 }
671
672 buffer = wbsd_kmap_sg(host);
673 }
674 }
675 }
676
677 wbsd_kunmap_sg(host);
678
679 /*
680 * The controller stops sending interrupts for
681 * 'FIFO empty' under certain conditions. So we
682 * need to be a bit more pro-active.
683 */
684 tasklet_schedule(&host->fifo_tasklet);
685 }
686
687 static void wbsd_prepare_data(struct wbsd_host* host, struct mmc_data* data)
688 {
689 u16 blksize;
690 u8 setup;
691 unsigned long dmaflags;
692
693 DBGF("blksz %04x blks %04x flags %08x\n",
694 1 << data->blksz_bits, data->blocks, data->flags);
695 DBGF("tsac %d ms nsac %d clk\n",
696 data->timeout_ns / 1000000, data->timeout_clks);
697
698 /*
699 * Calculate size.
700 */
701 host->size = data->blocks << data->blksz_bits;
702
703 /*
704 * Check timeout values for overflow.
705 * (Yes, some cards cause this value to overflow).
706 */
707 if (data->timeout_ns > 127000000)
708 wbsd_write_index(host, WBSD_IDX_TAAC, 127);
709 else
710 wbsd_write_index(host, WBSD_IDX_TAAC, data->timeout_ns/1000000);
711
712 if (data->timeout_clks > 255)
713 wbsd_write_index(host, WBSD_IDX_NSAC, 255);
714 else
715 wbsd_write_index(host, WBSD_IDX_NSAC, data->timeout_clks);
716
717 /*
718 * Inform the chip of how large blocks will be
719 * sent. It needs this to determine when to
720 * calculate CRC.
721 *
722 * Space for CRC must be included in the size.
723 * Two bytes are needed for each data line.
724 */
725 if (host->bus_width == MMC_BUS_WIDTH_1)
726 {
727 blksize = (1 << data->blksz_bits) + 2;
728
729 wbsd_write_index(host, WBSD_IDX_PBSMSB, (blksize >> 4) & 0xF0);
730 wbsd_write_index(host, WBSD_IDX_PBSLSB, blksize & 0xFF);
731 }
732 else if (host->bus_width == MMC_BUS_WIDTH_4)
733 {
734 blksize = (1 << data->blksz_bits) + 2 * 4;
735
736 wbsd_write_index(host, WBSD_IDX_PBSMSB, ((blksize >> 4) & 0xF0)
737 | WBSD_DATA_WIDTH);
738 wbsd_write_index(host, WBSD_IDX_PBSLSB, blksize & 0xFF);
739 }
740 else
741 {
742 data->error = MMC_ERR_INVALID;
743 return;
744 }
745
746 /*
747 * Clear the FIFO. This is needed even for DMA
748 * transfers since the chip still uses the FIFO
749 * internally.
750 */
751 setup = wbsd_read_index(host, WBSD_IDX_SETUP);
752 setup |= WBSD_FIFO_RESET;
753 wbsd_write_index(host, WBSD_IDX_SETUP, setup);
754
755 /*
756 * DMA transfer?
757 */
758 if (host->dma >= 0)
759 {
760 /*
761 * The buffer for DMA is only 64 kB.
762 */
763 BUG_ON(host->size > 0x10000);
764 if (host->size > 0x10000)
765 {
766 data->error = MMC_ERR_INVALID;
767 return;
768 }
769
770 /*
771 * Transfer data from the SG list to
772 * the DMA buffer.
773 */
774 if (data->flags & MMC_DATA_WRITE)
775 wbsd_sg_to_dma(host, data);
776
777 /*
778 * Initialise the ISA DMA controller.
779 */
780 dmaflags = claim_dma_lock();
781 disable_dma(host->dma);
782 clear_dma_ff(host->dma);
783 if (data->flags & MMC_DATA_READ)
784 set_dma_mode(host->dma, DMA_MODE_READ & ~0x40);
785 else
786 set_dma_mode(host->dma, DMA_MODE_WRITE & ~0x40);
787 set_dma_addr(host->dma, host->dma_addr);
788 set_dma_count(host->dma, host->size);
789
790 enable_dma(host->dma);
791 release_dma_lock(dmaflags);
792
793 /*
794 * Enable DMA on the host.
795 */
796 wbsd_write_index(host, WBSD_IDX_DMA, WBSD_DMA_ENABLE);
797 }
798 else
799 {
800 /*
801 * This flag is used to keep printk
802 * output to a minimum.
803 */
804 host->firsterr = 1;
805
806 /*
807 * Initialise the SG list.
808 */
809 wbsd_init_sg(host, data);
810
811 /*
812 * Turn off DMA.
813 */
814 wbsd_write_index(host, WBSD_IDX_DMA, 0);
815
816 /*
817 * Set up FIFO threshold levels (and fill
818 * buffer if doing a write).
819 */
820 if (data->flags & MMC_DATA_READ)
821 {
822 wbsd_write_index(host, WBSD_IDX_FIFOEN,
823 WBSD_FIFOEN_FULL | 8);
824 }
825 else
826 {
827 wbsd_write_index(host, WBSD_IDX_FIFOEN,
828 WBSD_FIFOEN_EMPTY | 8);
829 wbsd_fill_fifo(host);
830 }
831 }
832
833 data->error = MMC_ERR_NONE;
834 }
835
836 static void wbsd_finish_data(struct wbsd_host* host, struct mmc_data* data)
837 {
838 unsigned long dmaflags;
839 int count;
840 u8 status;
841
842 WARN_ON(host->mrq == NULL);
843
844 /*
845 * Send a stop command if needed.
846 */
847 if (data->stop)
848 wbsd_send_command(host, data->stop);
849
850 /*
851 * Wait for the controller to leave data
852 * transfer state.
853 */
854 do
855 {
856 status = wbsd_read_index(host, WBSD_IDX_STATUS);
857 } while (status & (WBSD_BLOCK_READ | WBSD_BLOCK_WRITE));
858
859 /*
860 * DMA transfer?
861 */
862 if (host->dma >= 0)
863 {
864 /*
865 * Disable DMA on the host.
866 */
867 wbsd_write_index(host, WBSD_IDX_DMA, 0);
868
869 /*
870 * Turn of ISA DMA controller.
871 */
872 dmaflags = claim_dma_lock();
873 disable_dma(host->dma);
874 clear_dma_ff(host->dma);
875 count = get_dma_residue(host->dma);
876 release_dma_lock(dmaflags);
877
878 /*
879 * Any leftover data?
880 */
881 if (count)
882 {
883 printk(KERN_ERR DRIVER_NAME ": Incomplete DMA "
884 "transfer. %d bytes left.\n", count);
885
886 data->error = MMC_ERR_FAILED;
887 }
888 else
889 {
890 /*
891 * Transfer data from DMA buffer to
892 * SG list.
893 */
894 if (data->flags & MMC_DATA_READ)
895 wbsd_dma_to_sg(host, data);
896
897 data->bytes_xfered = host->size;
898 }
899 }
900
901 DBGF("Ending data transfer (%d bytes)\n", data->bytes_xfered);
902
903 wbsd_request_end(host, host->mrq);
904 }
905
906 /*****************************************************************************\
907 * *
908 * MMC layer callbacks *
909 * *
910 \*****************************************************************************/
911
912 static void wbsd_request(struct mmc_host* mmc, struct mmc_request* mrq)
913 {
914 struct wbsd_host* host = mmc_priv(mmc);
915 struct mmc_command* cmd;
916
917 /*
918 * Disable tasklets to avoid a deadlock.
919 */
920 spin_lock_bh(&host->lock);
921
922 BUG_ON(host->mrq != NULL);
923
924 cmd = mrq->cmd;
925
926 host->mrq = mrq;
927
928 /*
929 * If there is no card in the slot then
930 * timeout immediatly.
931 */
932 if (!(host->flags & WBSD_FCARD_PRESENT))
933 {
934 cmd->error = MMC_ERR_TIMEOUT;
935 goto done;
936 }
937
938 /*
939 * Does the request include data?
940 */
941 if (cmd->data)
942 {
943 wbsd_prepare_data(host, cmd->data);
944
945 if (cmd->data->error != MMC_ERR_NONE)
946 goto done;
947 }
948
949 wbsd_send_command(host, cmd);
950
951 /*
952 * If this is a data transfer the request
953 * will be finished after the data has
954 * transfered.
955 */
956 if (cmd->data && (cmd->error == MMC_ERR_NONE))
957 {
958 /*
959 * Dirty fix for hardware bug.
960 */
961 if (host->dma == -1)
962 tasklet_schedule(&host->fifo_tasklet);
963
964 spin_unlock_bh(&host->lock);
965
966 return;
967 }
968
969 done:
970 wbsd_request_end(host, mrq);
971
972 spin_unlock_bh(&host->lock);
973 }
974
975 static void wbsd_set_ios(struct mmc_host* mmc, struct mmc_ios* ios)
976 {
977 struct wbsd_host* host = mmc_priv(mmc);
978 u8 clk, setup, pwr;
979
980 DBGF("clock %uHz busmode %u powermode %u cs %u Vdd %u width %u\n",
981 ios->clock, ios->bus_mode, ios->power_mode, ios->chip_select,
982 ios->vdd, ios->bus_width);
983
984 spin_lock_bh(&host->lock);
985
986 /*
987 * Reset the chip on each power off.
988 * Should clear out any weird states.
989 */
990 if (ios->power_mode == MMC_POWER_OFF)
991 wbsd_init_device(host);
992
993 if (ios->clock >= 24000000)
994 clk = WBSD_CLK_24M;
995 else if (ios->clock >= 16000000)
996 clk = WBSD_CLK_16M;
997 else if (ios->clock >= 12000000)
998 clk = WBSD_CLK_12M;
999 else
1000 clk = WBSD_CLK_375K;
1001
1002 /*
1003 * Only write to the clock register when
1004 * there is an actual change.
1005 */
1006 if (clk != host->clk)
1007 {
1008 wbsd_write_index(host, WBSD_IDX_CLK, clk);
1009 host->clk = clk;
1010 }
1011
1012 /*
1013 * Power up card.
1014 */
1015 if (ios->power_mode != MMC_POWER_OFF)
1016 {
1017 pwr = inb(host->base + WBSD_CSR);
1018 pwr &= ~WBSD_POWER_N;
1019 outb(pwr, host->base + WBSD_CSR);
1020 }
1021
1022 /*
1023 * MMC cards need to have pin 1 high during init.
1024 * It wreaks havoc with the card detection though so
1025 * that needs to be disabled.
1026 */
1027 setup = wbsd_read_index(host, WBSD_IDX_SETUP);
1028 if (ios->chip_select == MMC_CS_HIGH)
1029 {
1030 BUG_ON(ios->bus_width != MMC_BUS_WIDTH_1);
1031 setup |= WBSD_DAT3_H;
1032 host->flags |= WBSD_FIGNORE_DETECT;
1033 }
1034 else
1035 {
1036 if (setup & WBSD_DAT3_H)
1037 {
1038 setup &= ~WBSD_DAT3_H;
1039
1040 /*
1041 * We cannot resume card detection immediatly
1042 * because of capacitance and delays in the chip.
1043 */
1044 mod_timer(&host->ignore_timer, jiffies + HZ/100);
1045 }
1046 }
1047 wbsd_write_index(host, WBSD_IDX_SETUP, setup);
1048
1049 /*
1050 * Store bus width for later. Will be used when
1051 * setting up the data transfer.
1052 */
1053 host->bus_width = ios->bus_width;
1054
1055 spin_unlock_bh(&host->lock);
1056 }
1057
1058 static int wbsd_get_ro(struct mmc_host* mmc)
1059 {
1060 struct wbsd_host* host = mmc_priv(mmc);
1061 u8 csr;
1062
1063 spin_lock_bh(&host->lock);
1064
1065 csr = inb(host->base + WBSD_CSR);
1066 csr |= WBSD_MSLED;
1067 outb(csr, host->base + WBSD_CSR);
1068
1069 mdelay(1);
1070
1071 csr = inb(host->base + WBSD_CSR);
1072 csr &= ~WBSD_MSLED;
1073 outb(csr, host->base + WBSD_CSR);
1074
1075 spin_unlock_bh(&host->lock);
1076
1077 return csr & WBSD_WRPT;
1078 }
1079
1080 static struct mmc_host_ops wbsd_ops = {
1081 .request = wbsd_request,
1082 .set_ios = wbsd_set_ios,
1083 .get_ro = wbsd_get_ro,
1084 };
1085
1086 /*****************************************************************************\
1087 * *
1088 * Interrupt handling *
1089 * *
1090 \*****************************************************************************/
1091
1092 /*
1093 * Helper function to reset detection ignore
1094 */
1095
1096 static void wbsd_reset_ignore(unsigned long data)
1097 {
1098 struct wbsd_host *host = (struct wbsd_host*)data;
1099
1100 BUG_ON(host == NULL);
1101
1102 DBG("Resetting card detection ignore\n");
1103
1104 spin_lock_bh(&host->lock);
1105
1106 host->flags &= ~WBSD_FIGNORE_DETECT;
1107
1108 /*
1109 * Card status might have changed during the
1110 * blackout.
1111 */
1112 tasklet_schedule(&host->card_tasklet);
1113
1114 spin_unlock_bh(&host->lock);
1115 }
1116
1117 /*
1118 * Tasklets
1119 */
1120
1121 static inline struct mmc_data* wbsd_get_data(struct wbsd_host* host)
1122 {
1123 WARN_ON(!host->mrq);
1124 if (!host->mrq)
1125 return NULL;
1126
1127 WARN_ON(!host->mrq->cmd);
1128 if (!host->mrq->cmd)
1129 return NULL;
1130
1131 WARN_ON(!host->mrq->cmd->data);
1132 if (!host->mrq->cmd->data)
1133 return NULL;
1134
1135 return host->mrq->cmd->data;
1136 }
1137
1138 static void wbsd_tasklet_card(unsigned long param)
1139 {
1140 struct wbsd_host* host = (struct wbsd_host*)param;
1141 u8 csr;
1142 int delay = -1;
1143
1144 spin_lock(&host->lock);
1145
1146 if (host->flags & WBSD_FIGNORE_DETECT)
1147 {
1148 spin_unlock(&host->lock);
1149 return;
1150 }
1151
1152 csr = inb(host->base + WBSD_CSR);
1153 WARN_ON(csr == 0xff);
1154
1155 if (csr & WBSD_CARDPRESENT)
1156 {
1157 if (!(host->flags & WBSD_FCARD_PRESENT))
1158 {
1159 DBG("Card inserted\n");
1160 host->flags |= WBSD_FCARD_PRESENT;
1161
1162 delay = 500;
1163 }
1164 }
1165 else if (host->flags & WBSD_FCARD_PRESENT)
1166 {
1167 DBG("Card removed\n");
1168 host->flags &= ~WBSD_FCARD_PRESENT;
1169
1170 if (host->mrq)
1171 {
1172 printk(KERN_ERR DRIVER_NAME
1173 ": Card removed during transfer!\n");
1174 wbsd_reset(host);
1175
1176 host->mrq->cmd->error = MMC_ERR_FAILED;
1177 tasklet_schedule(&host->finish_tasklet);
1178 }
1179
1180 delay = 0;
1181 }
1182
1183 /*
1184 * Unlock first since we might get a call back.
1185 */
1186
1187 spin_unlock(&host->lock);
1188
1189 if (delay != -1)
1190 mmc_detect_change(host->mmc, msecs_to_jiffies(delay));
1191 }
1192
1193 static void wbsd_tasklet_fifo(unsigned long param)
1194 {
1195 struct wbsd_host* host = (struct wbsd_host*)param;
1196 struct mmc_data* data;
1197
1198 spin_lock(&host->lock);
1199
1200 if (!host->mrq)
1201 goto end;
1202
1203 data = wbsd_get_data(host);
1204 if (!data)
1205 goto end;
1206
1207 if (data->flags & MMC_DATA_WRITE)
1208 wbsd_fill_fifo(host);
1209 else
1210 wbsd_empty_fifo(host);
1211
1212 /*
1213 * Done?
1214 */
1215 if (host->size == data->bytes_xfered)
1216 {
1217 wbsd_write_index(host, WBSD_IDX_FIFOEN, 0);
1218 tasklet_schedule(&host->finish_tasklet);
1219 }
1220
1221 end:
1222 spin_unlock(&host->lock);
1223 }
1224
1225 static void wbsd_tasklet_crc(unsigned long param)
1226 {
1227 struct wbsd_host* host = (struct wbsd_host*)param;
1228 struct mmc_data* data;
1229
1230 spin_lock(&host->lock);
1231
1232 if (!host->mrq)
1233 goto end;
1234
1235 data = wbsd_get_data(host);
1236 if (!data)
1237 goto end;
1238
1239 DBGF("CRC error\n");
1240
1241 data->error = MMC_ERR_BADCRC;
1242
1243 tasklet_schedule(&host->finish_tasklet);
1244
1245 end:
1246 spin_unlock(&host->lock);
1247 }
1248
1249 static void wbsd_tasklet_timeout(unsigned long param)
1250 {
1251 struct wbsd_host* host = (struct wbsd_host*)param;
1252 struct mmc_data* data;
1253
1254 spin_lock(&host->lock);
1255
1256 if (!host->mrq)
1257 goto end;
1258
1259 data = wbsd_get_data(host);
1260 if (!data)
1261 goto end;
1262
1263 DBGF("Timeout\n");
1264
1265 data->error = MMC_ERR_TIMEOUT;
1266
1267 tasklet_schedule(&host->finish_tasklet);
1268
1269 end:
1270 spin_unlock(&host->lock);
1271 }
1272
1273 static void wbsd_tasklet_finish(unsigned long param)
1274 {
1275 struct wbsd_host* host = (struct wbsd_host*)param;
1276 struct mmc_data* data;
1277
1278 spin_lock(&host->lock);
1279
1280 WARN_ON(!host->mrq);
1281 if (!host->mrq)
1282 goto end;
1283
1284 data = wbsd_get_data(host);
1285 if (!data)
1286 goto end;
1287
1288 wbsd_finish_data(host, data);
1289
1290 end:
1291 spin_unlock(&host->lock);
1292 }
1293
1294 static void wbsd_tasklet_block(unsigned long param)
1295 {
1296 struct wbsd_host* host = (struct wbsd_host*)param;
1297 struct mmc_data* data;
1298
1299 spin_lock(&host->lock);
1300
1301 if ((wbsd_read_index(host, WBSD_IDX_CRCSTATUS) & WBSD_CRC_MASK) !=
1302 WBSD_CRC_OK)
1303 {
1304 data = wbsd_get_data(host);
1305 if (!data)
1306 goto end;
1307
1308 DBGF("CRC error\n");
1309
1310 data->error = MMC_ERR_BADCRC;
1311
1312 tasklet_schedule(&host->finish_tasklet);
1313 }
1314
1315 end:
1316 spin_unlock(&host->lock);
1317 }
1318
1319 /*
1320 * Interrupt handling
1321 */
1322
1323 static irqreturn_t wbsd_irq(int irq, void *dev_id, struct pt_regs *regs)
1324 {
1325 struct wbsd_host* host = dev_id;
1326 int isr;
1327
1328 isr = inb(host->base + WBSD_ISR);
1329
1330 /*
1331 * Was it actually our hardware that caused the interrupt?
1332 */
1333 if (isr == 0xff || isr == 0x00)
1334 return IRQ_NONE;
1335
1336 host->isr |= isr;
1337
1338 /*
1339 * Schedule tasklets as needed.
1340 */
1341 if (isr & WBSD_INT_CARD)
1342 tasklet_schedule(&host->card_tasklet);
1343 if (isr & WBSD_INT_FIFO_THRE)
1344 tasklet_schedule(&host->fifo_tasklet);
1345 if (isr & WBSD_INT_CRC)
1346 tasklet_hi_schedule(&host->crc_tasklet);
1347 if (isr & WBSD_INT_TIMEOUT)
1348 tasklet_hi_schedule(&host->timeout_tasklet);
1349 if (isr & WBSD_INT_BUSYEND)
1350 tasklet_hi_schedule(&host->block_tasklet);
1351 if (isr & WBSD_INT_TC)
1352 tasklet_schedule(&host->finish_tasklet);
1353
1354 return IRQ_HANDLED;
1355 }
1356
1357 /*****************************************************************************\
1358 * *
1359 * Device initialisation and shutdown *
1360 * *
1361 \*****************************************************************************/
1362
1363 /*
1364 * Allocate/free MMC structure.
1365 */
1366
1367 static int __devinit wbsd_alloc_mmc(struct device* dev)
1368 {
1369 struct mmc_host* mmc;
1370 struct wbsd_host* host;
1371
1372 /*
1373 * Allocate MMC structure.
1374 */
1375 mmc = mmc_alloc_host(sizeof(struct wbsd_host), dev);
1376 if (!mmc)
1377 return -ENOMEM;
1378
1379 host = mmc_priv(mmc);
1380 host->mmc = mmc;
1381
1382 host->dma = -1;
1383
1384 /*
1385 * Set host parameters.
1386 */
1387 mmc->ops = &wbsd_ops;
1388 mmc->f_min = 375000;
1389 mmc->f_max = 24000000;
1390 mmc->ocr_avail = MMC_VDD_32_33|MMC_VDD_33_34;
1391 mmc->caps = MMC_CAP_4_BIT_DATA;
1392
1393 spin_lock_init(&host->lock);
1394
1395 /*
1396 * Set up timers
1397 */
1398 init_timer(&host->ignore_timer);
1399 host->ignore_timer.data = (unsigned long)host;
1400 host->ignore_timer.function = wbsd_reset_ignore;
1401
1402 /*
1403 * Maximum number of segments. Worst case is one sector per segment
1404 * so this will be 64kB/512.
1405 */
1406 mmc->max_hw_segs = 128;
1407 mmc->max_phys_segs = 128;
1408
1409 /*
1410 * Maximum number of sectors in one transfer. Also limited by 64kB
1411 * buffer.
1412 */
1413 mmc->max_sectors = 128;
1414
1415 /*
1416 * Maximum segment size. Could be one segment with the maximum number
1417 * of segments.
1418 */
1419 mmc->max_seg_size = mmc->max_sectors * 512;
1420
1421 dev_set_drvdata(dev, mmc);
1422
1423 return 0;
1424 }
1425
1426 static void __devexit wbsd_free_mmc(struct device* dev)
1427 {
1428 struct mmc_host* mmc;
1429 struct wbsd_host* host;
1430
1431 mmc = dev_get_drvdata(dev);
1432 if (!mmc)
1433 return;
1434
1435 host = mmc_priv(mmc);
1436 BUG_ON(host == NULL);
1437
1438 del_timer_sync(&host->ignore_timer);
1439
1440 mmc_free_host(mmc);
1441
1442 dev_set_drvdata(dev, NULL);
1443 }
1444
1445 /*
1446 * Scan for known chip id:s
1447 */
1448
1449 static int __devinit wbsd_scan(struct wbsd_host* host)
1450 {
1451 int i, j, k;
1452 int id;
1453
1454 /*
1455 * Iterate through all ports, all codes to
1456 * find hardware that is in our known list.
1457 */
1458 for (i = 0;i < sizeof(config_ports)/sizeof(int);i++)
1459 {
1460 if (!request_region(config_ports[i], 2, DRIVER_NAME))
1461 continue;
1462
1463 for (j = 0;j < sizeof(unlock_codes)/sizeof(int);j++)
1464 {
1465 id = 0xFFFF;
1466
1467 host->config = config_ports[i];
1468 host->unlock_code = unlock_codes[j];
1469
1470 wbsd_unlock_config(host);
1471
1472 outb(WBSD_CONF_ID_HI, config_ports[i]);
1473 id = inb(config_ports[i] + 1) << 8;
1474
1475 outb(WBSD_CONF_ID_LO, config_ports[i]);
1476 id |= inb(config_ports[i] + 1);
1477
1478 wbsd_lock_config(host);
1479
1480 for (k = 0;k < sizeof(valid_ids)/sizeof(int);k++)
1481 {
1482 if (id == valid_ids[k])
1483 {
1484 host->chip_id = id;
1485
1486 return 0;
1487 }
1488 }
1489
1490 if (id != 0xFFFF)
1491 {
1492 DBG("Unknown hardware (id %x) found at %x\n",
1493 id, config_ports[i]);
1494 }
1495 }
1496
1497 release_region(config_ports[i], 2);
1498 }
1499
1500 host->config = 0;
1501 host->unlock_code = 0;
1502
1503 return -ENODEV;
1504 }
1505
1506 /*
1507 * Allocate/free io port ranges
1508 */
1509
1510 static int __devinit wbsd_request_region(struct wbsd_host* host, int base)
1511 {
1512 if (io & 0x7)
1513 return -EINVAL;
1514
1515 if (!request_region(base, 8, DRIVER_NAME))
1516 return -EIO;
1517
1518 host->base = io;
1519
1520 return 0;
1521 }
1522
1523 static void __devexit wbsd_release_regions(struct wbsd_host* host)
1524 {
1525 if (host->base)
1526 release_region(host->base, 8);
1527
1528 host->base = 0;
1529
1530 if (host->config)
1531 release_region(host->config, 2);
1532
1533 host->config = 0;
1534 }
1535
1536 /*
1537 * Allocate/free DMA port and buffer
1538 */
1539
1540 static void __devinit wbsd_request_dma(struct wbsd_host* host, int dma)
1541 {
1542 if (dma < 0)
1543 return;
1544
1545 if (request_dma(dma, DRIVER_NAME))
1546 goto err;
1547
1548 /*
1549 * We need to allocate a special buffer in
1550 * order for ISA to be able to DMA to it.
1551 */
1552 host->dma_buffer = kmalloc(WBSD_DMA_SIZE,
1553 GFP_NOIO | GFP_DMA | __GFP_REPEAT | __GFP_NOWARN);
1554 if (!host->dma_buffer)
1555 goto free;
1556
1557 /*
1558 * Translate the address to a physical address.
1559 */
1560 host->dma_addr = dma_map_single(host->mmc->dev, host->dma_buffer,
1561 WBSD_DMA_SIZE, DMA_BIDIRECTIONAL);
1562
1563 /*
1564 * ISA DMA must be aligned on a 64k basis.
1565 */
1566 if ((host->dma_addr & 0xffff) != 0)
1567 goto kfree;
1568 /*
1569 * ISA cannot access memory above 16 MB.
1570 */
1571 else if (host->dma_addr >= 0x1000000)
1572 goto kfree;
1573
1574 host->dma = dma;
1575
1576 return;
1577
1578 kfree:
1579 /*
1580 * If we've gotten here then there is some kind of alignment bug
1581 */
1582 BUG_ON(1);
1583
1584 dma_unmap_single(host->mmc->dev, host->dma_addr, WBSD_DMA_SIZE,
1585 DMA_BIDIRECTIONAL);
1586 host->dma_addr = (dma_addr_t)NULL;
1587
1588 kfree(host->dma_buffer);
1589 host->dma_buffer = NULL;
1590
1591 free:
1592 free_dma(dma);
1593
1594 err:
1595 printk(KERN_WARNING DRIVER_NAME ": Unable to allocate DMA %d. "
1596 "Falling back on FIFO.\n", dma);
1597 }
1598
1599 static void __devexit wbsd_release_dma(struct wbsd_host* host)
1600 {
1601 if (host->dma_addr)
1602 dma_unmap_single(host->mmc->dev, host->dma_addr, WBSD_DMA_SIZE,
1603 DMA_BIDIRECTIONAL);
1604 if (host->dma_buffer)
1605 kfree(host->dma_buffer);
1606 if (host->dma >= 0)
1607 free_dma(host->dma);
1608
1609 host->dma = -1;
1610 host->dma_buffer = NULL;
1611 host->dma_addr = (dma_addr_t)NULL;
1612 }
1613
1614 /*
1615 * Allocate/free IRQ.
1616 */
1617
1618 static int __devinit wbsd_request_irq(struct wbsd_host* host, int irq)
1619 {
1620 int ret;
1621
1622 /*
1623 * Allocate interrupt.
1624 */
1625
1626 ret = request_irq(irq, wbsd_irq, SA_SHIRQ, DRIVER_NAME, host);
1627 if (ret)
1628 return ret;
1629
1630 host->irq = irq;
1631
1632 /*
1633 * Set up tasklets.
1634 */
1635 tasklet_init(&host->card_tasklet, wbsd_tasklet_card, (unsigned long)host);
1636 tasklet_init(&host->fifo_tasklet, wbsd_tasklet_fifo, (unsigned long)host);
1637 tasklet_init(&host->crc_tasklet, wbsd_tasklet_crc, (unsigned long)host);
1638 tasklet_init(&host->timeout_tasklet, wbsd_tasklet_timeout, (unsigned long)host);
1639 tasklet_init(&host->finish_tasklet, wbsd_tasklet_finish, (unsigned long)host);
1640 tasklet_init(&host->block_tasklet, wbsd_tasklet_block, (unsigned long)host);
1641
1642 return 0;
1643 }
1644
1645 static void __devexit wbsd_release_irq(struct wbsd_host* host)
1646 {
1647 if (!host->irq)
1648 return;
1649
1650 free_irq(host->irq, host);
1651
1652 host->irq = 0;
1653
1654 tasklet_kill(&host->card_tasklet);
1655 tasklet_kill(&host->fifo_tasklet);
1656 tasklet_kill(&host->crc_tasklet);
1657 tasklet_kill(&host->timeout_tasklet);
1658 tasklet_kill(&host->finish_tasklet);
1659 tasklet_kill(&host->block_tasklet);
1660 }
1661
1662 /*
1663 * Allocate all resources for the host.
1664 */
1665
1666 static int __devinit wbsd_request_resources(struct wbsd_host* host,
1667 int base, int irq, int dma)
1668 {
1669 int ret;
1670
1671 /*
1672 * Allocate I/O ports.
1673 */
1674 ret = wbsd_request_region(host, base);
1675 if (ret)
1676 return ret;
1677
1678 /*
1679 * Allocate interrupt.
1680 */
1681 ret = wbsd_request_irq(host, irq);
1682 if (ret)
1683 return ret;
1684
1685 /*
1686 * Allocate DMA.
1687 */
1688 wbsd_request_dma(host, dma);
1689
1690 return 0;
1691 }
1692
1693 /*
1694 * Release all resources for the host.
1695 */
1696
1697 static void __devexit wbsd_release_resources(struct wbsd_host* host)
1698 {
1699 wbsd_release_dma(host);
1700 wbsd_release_irq(host);
1701 wbsd_release_regions(host);
1702 }
1703
1704 /*
1705 * Configure the resources the chip should use.
1706 */
1707
1708 static void wbsd_chip_config(struct wbsd_host* host)
1709 {
1710 wbsd_unlock_config(host);
1711
1712 /*
1713 * Reset the chip.
1714 */
1715 wbsd_write_config(host, WBSD_CONF_SWRST, 1);
1716 wbsd_write_config(host, WBSD_CONF_SWRST, 0);
1717
1718 /*
1719 * Select SD/MMC function.
1720 */
1721 wbsd_write_config(host, WBSD_CONF_DEVICE, DEVICE_SD);
1722
1723 /*
1724 * Set up card detection.
1725 */
1726 wbsd_write_config(host, WBSD_CONF_PINS, WBSD_PINS_DETECT_GP11);
1727
1728 /*
1729 * Configure chip
1730 */
1731 wbsd_write_config(host, WBSD_CONF_PORT_HI, host->base >> 8);
1732 wbsd_write_config(host, WBSD_CONF_PORT_LO, host->base & 0xff);
1733
1734 wbsd_write_config(host, WBSD_CONF_IRQ, host->irq);
1735
1736 if (host->dma >= 0)
1737 wbsd_write_config(host, WBSD_CONF_DRQ, host->dma);
1738
1739 /*
1740 * Enable and power up chip.
1741 */
1742 wbsd_write_config(host, WBSD_CONF_ENABLE, 1);
1743 wbsd_write_config(host, WBSD_CONF_POWER, 0x20);
1744
1745 wbsd_lock_config(host);
1746 }
1747
1748 /*
1749 * Check that configured resources are correct.
1750 */
1751
1752 static int wbsd_chip_validate(struct wbsd_host* host)
1753 {
1754 int base, irq, dma;
1755
1756 wbsd_unlock_config(host);
1757
1758 /*
1759 * Select SD/MMC function.
1760 */
1761 wbsd_write_config(host, WBSD_CONF_DEVICE, DEVICE_SD);
1762
1763 /*
1764 * Read configuration.
1765 */
1766 base = wbsd_read_config(host, WBSD_CONF_PORT_HI) << 8;
1767 base |= wbsd_read_config(host, WBSD_CONF_PORT_LO);
1768
1769 irq = wbsd_read_config(host, WBSD_CONF_IRQ);
1770
1771 dma = wbsd_read_config(host, WBSD_CONF_DRQ);
1772
1773 wbsd_lock_config(host);
1774
1775 /*
1776 * Validate against given configuration.
1777 */
1778 if (base != host->base)
1779 return 0;
1780 if (irq != host->irq)
1781 return 0;
1782 if ((dma != host->dma) && (host->dma != -1))
1783 return 0;
1784
1785 return 1;
1786 }
1787
1788 /*
1789 * Powers down the SD function
1790 */
1791
1792 static void wbsd_chip_poweroff(struct wbsd_host* host)
1793 {
1794 wbsd_unlock_config(host);
1795
1796 wbsd_write_config(host, WBSD_CONF_DEVICE, DEVICE_SD);
1797 wbsd_write_config(host, WBSD_CONF_ENABLE, 0);
1798
1799 wbsd_lock_config(host);
1800 }
1801
1802 /*****************************************************************************\
1803 * *
1804 * Devices setup and shutdown *
1805 * *
1806 \*****************************************************************************/
1807
1808 static int __devinit wbsd_init(struct device* dev, int base, int irq, int dma,
1809 int pnp)
1810 {
1811 struct wbsd_host* host = NULL;
1812 struct mmc_host* mmc = NULL;
1813 int ret;
1814
1815 ret = wbsd_alloc_mmc(dev);
1816 if (ret)
1817 return ret;
1818
1819 mmc = dev_get_drvdata(dev);
1820 host = mmc_priv(mmc);
1821
1822 /*
1823 * Scan for hardware.
1824 */
1825 ret = wbsd_scan(host);
1826 if (ret)
1827 {
1828 if (pnp && (ret == -ENODEV))
1829 {
1830 printk(KERN_WARNING DRIVER_NAME
1831 ": Unable to confirm device presence. You may "
1832 "experience lock-ups.\n");
1833 }
1834 else
1835 {
1836 wbsd_free_mmc(dev);
1837 return ret;
1838 }
1839 }
1840
1841 /*
1842 * Request resources.
1843 */
1844 ret = wbsd_request_resources(host, io, irq, dma);
1845 if (ret)
1846 {
1847 wbsd_release_resources(host);
1848 wbsd_free_mmc(dev);
1849 return ret;
1850 }
1851
1852 /*
1853 * See if chip needs to be configured.
1854 */
1855 if (pnp && (host->config != 0))
1856 {
1857 if (!wbsd_chip_validate(host))
1858 {
1859 printk(KERN_WARNING DRIVER_NAME
1860 ": PnP active but chip not configured! "
1861 "You probably have a buggy BIOS. "
1862 "Configuring chip manually.\n");
1863 wbsd_chip_config(host);
1864 }
1865 }
1866 else
1867 wbsd_chip_config(host);
1868
1869 /*
1870 * Power Management stuff. No idea how this works.
1871 * Not tested.
1872 */
1873 #ifdef CONFIG_PM
1874 if (host->config)
1875 {
1876 wbsd_unlock_config(host);
1877 wbsd_write_config(host, WBSD_CONF_PME, 0xA0);
1878 wbsd_lock_config(host);
1879 }
1880 #endif
1881 /*
1882 * Allow device to initialise itself properly.
1883 */
1884 mdelay(5);
1885
1886 /*
1887 * Reset the chip into a known state.
1888 */
1889 wbsd_init_device(host);
1890
1891 mmc_add_host(mmc);
1892
1893 printk(KERN_INFO "%s: W83L51xD", mmc_hostname(mmc));
1894 if (host->chip_id != 0)
1895 printk(" id %x", (int)host->chip_id);
1896 printk(" at 0x%x irq %d", (int)host->base, (int)host->irq);
1897 if (host->dma >= 0)
1898 printk(" dma %d", (int)host->dma);
1899 else
1900 printk(" FIFO");
1901 if (pnp)
1902 printk(" PnP");
1903 printk("\n");
1904
1905 return 0;
1906 }
1907
1908 static void __devexit wbsd_shutdown(struct device* dev, int pnp)
1909 {
1910 struct mmc_host* mmc = dev_get_drvdata(dev);
1911 struct wbsd_host* host;
1912
1913 if (!mmc)
1914 return;
1915
1916 host = mmc_priv(mmc);
1917
1918 mmc_remove_host(mmc);
1919
1920 /*
1921 * Power down the SD/MMC function.
1922 */
1923 if (!pnp)
1924 wbsd_chip_poweroff(host);
1925
1926 wbsd_release_resources(host);
1927
1928 wbsd_free_mmc(dev);
1929 }
1930
1931 /*
1932 * Non-PnP
1933 */
1934
1935 static int __devinit wbsd_probe(struct device* dev)
1936 {
1937 return wbsd_init(dev, io, irq, dma, 0);
1938 }
1939
1940 static int __devexit wbsd_remove(struct device* dev)
1941 {
1942 wbsd_shutdown(dev, 0);
1943
1944 return 0;
1945 }
1946
1947 /*
1948 * PnP
1949 */
1950
1951 #ifdef CONFIG_PNP
1952
1953 static int __devinit
1954 wbsd_pnp_probe(struct pnp_dev * pnpdev, const struct pnp_device_id *dev_id)
1955 {
1956 int io, irq, dma;
1957
1958 /*
1959 * Get resources from PnP layer.
1960 */
1961 io = pnp_port_start(pnpdev, 0);
1962 irq = pnp_irq(pnpdev, 0);
1963 if (pnp_dma_valid(pnpdev, 0))
1964 dma = pnp_dma(pnpdev, 0);
1965 else
1966 dma = -1;
1967
1968 DBGF("PnP resources: port %3x irq %d dma %d\n", io, irq, dma);
1969
1970 return wbsd_init(&pnpdev->dev, io, irq, dma, 1);
1971 }
1972
1973 static void __devexit wbsd_pnp_remove(struct pnp_dev * dev)
1974 {
1975 wbsd_shutdown(&dev->dev, 1);
1976 }
1977
1978 #endif /* CONFIG_PNP */
1979
1980 /*
1981 * Power management
1982 */
1983
1984 #ifdef CONFIG_PM
1985
1986 static int wbsd_suspend(struct device *dev, pm_message_t state)
1987 {
1988 struct mmc_host *mmc = dev_get_drvdata(dev);
1989 struct wbsd_host *host;
1990 int ret;
1991
1992 if (!mmc)
1993 return 0;
1994
1995 DBG("Suspending...\n");
1996
1997 ret = mmc_suspend_host(mmc, state);
1998 if (!ret)
1999 return ret;
2000
2001 host = mmc_priv(mmc);
2002
2003 wbsd_chip_poweroff(host);
2004
2005 return 0;
2006 }
2007
2008 static int wbsd_resume(struct device *dev)
2009 {
2010 struct mmc_host *mmc = dev_get_drvdata(dev);
2011 struct wbsd_host *host;
2012
2013 if (!mmc)
2014 return 0;
2015
2016 DBG("Resuming...\n");
2017
2018 host = mmc_priv(mmc);
2019
2020 wbsd_chip_config(host);
2021
2022 /*
2023 * Allow device to initialise itself properly.
2024 */
2025 mdelay(5);
2026
2027 wbsd_init_device(host);
2028
2029 return mmc_resume_host(mmc);
2030 }
2031
2032 #else /* CONFIG_PM */
2033
2034 #define wbsd_suspend NULL
2035 #define wbsd_resume NULL
2036
2037 #endif /* CONFIG_PM */
2038
2039 static struct platform_device *wbsd_device;
2040
2041 static struct device_driver wbsd_driver = {
2042 .name = DRIVER_NAME,
2043 .bus = &platform_bus_type,
2044 .probe = wbsd_probe,
2045 .remove = wbsd_remove,
2046
2047 .suspend = wbsd_suspend,
2048 .resume = wbsd_resume,
2049 };
2050
2051 #ifdef CONFIG_PNP
2052
2053 static struct pnp_driver wbsd_pnp_driver = {
2054 .name = DRIVER_NAME,
2055 .id_table = pnp_dev_table,
2056 .probe = wbsd_pnp_probe,
2057 .remove = wbsd_pnp_remove,
2058 };
2059
2060 #endif /* CONFIG_PNP */
2061
2062 /*
2063 * Module loading/unloading
2064 */
2065
2066 static int __init wbsd_drv_init(void)
2067 {
2068 int result;
2069
2070 printk(KERN_INFO DRIVER_NAME
2071 ": Winbond W83L51xD SD/MMC card interface driver, "
2072 DRIVER_VERSION "\n");
2073 printk(KERN_INFO DRIVER_NAME ": Copyright(c) Pierre Ossman\n");
2074
2075 #ifdef CONFIG_PNP
2076
2077 if (!nopnp)
2078 {
2079 result = pnp_register_driver(&wbsd_pnp_driver);
2080 if (result < 0)
2081 return result;
2082 }
2083
2084 #endif /* CONFIG_PNP */
2085
2086 if (nopnp)
2087 {
2088 result = driver_register(&wbsd_driver);
2089 if (result < 0)
2090 return result;
2091
2092 wbsd_device = platform_device_register_simple(DRIVER_NAME, -1,
2093 NULL, 0);
2094 if (IS_ERR(wbsd_device))
2095 return PTR_ERR(wbsd_device);
2096 }
2097
2098 return 0;
2099 }
2100
2101 static void __exit wbsd_drv_exit(void)
2102 {
2103 #ifdef CONFIG_PNP
2104
2105 if (!nopnp)
2106 pnp_unregister_driver(&wbsd_pnp_driver);
2107
2108 #endif /* CONFIG_PNP */
2109
2110 if (nopnp)
2111 {
2112 platform_device_unregister(wbsd_device);
2113
2114 driver_unregister(&wbsd_driver);
2115 }
2116
2117 DBG("unloaded\n");
2118 }
2119
2120 module_init(wbsd_drv_init);
2121 module_exit(wbsd_drv_exit);
2122 #ifdef CONFIG_PNP
2123 module_param(nopnp, uint, 0444);
2124 #endif
2125 module_param(io, uint, 0444);
2126 module_param(irq, uint, 0444);
2127 module_param(dma, int, 0444);
2128
2129 MODULE_LICENSE("GPL");
2130 MODULE_DESCRIPTION("Winbond W83L51xD SD/MMC card interface driver");
2131 MODULE_VERSION(DRIVER_VERSION);
2132
2133 #ifdef CONFIG_PNP
2134 MODULE_PARM_DESC(nopnp, "Scan for device instead of relying on PNP. (default 0)");
2135 #endif
2136 MODULE_PARM_DESC(io, "I/O base to allocate. Must be 8 byte aligned. (default 0x248)");
2137 MODULE_PARM_DESC(irq, "IRQ to allocate. (default 6)");
2138 MODULE_PARM_DESC(dma, "DMA channel to allocate. -1 for no DMA. (default 2)");
This page took 0.077982 seconds and 5 git commands to generate.