]> asedeno.scripts.mit.edu Git - linux.git/blob - arch/arm/mach-rpc/dma.c
treewide: Replace GPLv2 boilerplate/reference with SPDX - rule 152
[linux.git] / arch / arm / mach-rpc / dma.c
1 /*
2  *  linux/arch/arm/mach-rpc/dma.c
3  *
4  *  Copyright (C) 1998 Russell King
5  *
6  * This program is free software; you can redistribute it and/or modify
7  * it under the terms of the GNU General Public License version 2 as
8  * published by the Free Software Foundation.
9  *
10  *  DMA functions specific to RiscPC architecture
11  */
12 #include <linux/mman.h>
13 #include <linux/init.h>
14 #include <linux/interrupt.h>
15 #include <linux/dma-mapping.h>
16 #include <linux/io.h>
17
18 #include <asm/page.h>
19 #include <asm/dma.h>
20 #include <asm/fiq.h>
21 #include <asm/irq.h>
22 #include <mach/hardware.h>
23 #include <linux/uaccess.h>
24
25 #include <asm/mach/dma.h>
26 #include <asm/hardware/iomd.h>
27
28 struct iomd_dma {
29         struct dma_struct       dma;
30         unsigned int            state;
31         unsigned long           base;           /* Controller base address */
32         int                     irq;            /* Controller IRQ */
33         struct scatterlist      cur_sg;         /* Current controller buffer */
34         dma_addr_t              dma_addr;
35         unsigned int            dma_len;
36 };
37
38 #if 0
39 typedef enum {
40         dma_size_8      = 1,
41         dma_size_16     = 2,
42         dma_size_32     = 4,
43         dma_size_128    = 16
44 } dma_size_t;
45 #endif
46
47 #define TRANSFER_SIZE   2
48
49 #define CURA    (0)
50 #define ENDA    (IOMD_IO0ENDA - IOMD_IO0CURA)
51 #define CURB    (IOMD_IO0CURB - IOMD_IO0CURA)
52 #define ENDB    (IOMD_IO0ENDB - IOMD_IO0CURA)
53 #define CR      (IOMD_IO0CR - IOMD_IO0CURA)
54 #define ST      (IOMD_IO0ST - IOMD_IO0CURA)
55
56 static void iomd_get_next_sg(struct scatterlist *sg, struct iomd_dma *idma)
57 {
58         unsigned long end, offset, flags = 0;
59
60         if (idma->dma.sg) {
61                 sg->dma_address = idma->dma_addr;
62                 offset = sg->dma_address & ~PAGE_MASK;
63
64                 end = offset + idma->dma_len;
65
66                 if (end > PAGE_SIZE)
67                         end = PAGE_SIZE;
68
69                 if (offset + TRANSFER_SIZE >= end)
70                         flags |= DMA_END_L;
71
72                 sg->length = end - TRANSFER_SIZE;
73
74                 idma->dma_len -= end - offset;
75                 idma->dma_addr += end - offset;
76
77                 if (idma->dma_len == 0) {
78                         if (idma->dma.sgcount > 1) {
79                                 idma->dma.sg = sg_next(idma->dma.sg);
80                                 idma->dma_addr = idma->dma.sg->dma_address;
81                                 idma->dma_len = idma->dma.sg->length;
82                                 idma->dma.sgcount--;
83                         } else {
84                                 idma->dma.sg = NULL;
85                                 flags |= DMA_END_S;
86                         }
87                 }
88         } else {
89                 flags = DMA_END_S | DMA_END_L;
90                 sg->dma_address = 0;
91                 sg->length = 0;
92         }
93
94         sg->length |= flags;
95 }
96
97 static irqreturn_t iomd_dma_handle(int irq, void *dev_id)
98 {
99         struct iomd_dma *idma = dev_id;
100         unsigned long base = idma->base;
101
102         do {
103                 unsigned int status;
104
105                 status = iomd_readb(base + ST);
106                 if (!(status & DMA_ST_INT))
107                         return IRQ_HANDLED;
108
109                 if ((idma->state ^ status) & DMA_ST_AB)
110                         iomd_get_next_sg(&idma->cur_sg, idma);
111
112                 switch (status & (DMA_ST_OFL | DMA_ST_AB)) {
113                 case DMA_ST_OFL:                        /* OIA */
114                 case DMA_ST_AB:                         /* .IB */
115                         iomd_writel(idma->cur_sg.dma_address, base + CURA);
116                         iomd_writel(idma->cur_sg.length, base + ENDA);
117                         idma->state = DMA_ST_AB;
118                         break;
119
120                 case DMA_ST_OFL | DMA_ST_AB:            /* OIB */
121                 case 0:                                 /* .IA */
122                         iomd_writel(idma->cur_sg.dma_address, base + CURB);
123                         iomd_writel(idma->cur_sg.length, base + ENDB);
124                         idma->state = 0;
125                         break;
126                 }
127
128                 if (status & DMA_ST_OFL &&
129                     idma->cur_sg.length == (DMA_END_S|DMA_END_L))
130                         break;
131         } while (1);
132
133         idma->state = ~DMA_ST_AB;
134         disable_irq(irq);
135
136         return IRQ_HANDLED;
137 }
138
139 static int iomd_request_dma(unsigned int chan, dma_t *dma)
140 {
141         struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma);
142
143         return request_irq(idma->irq, iomd_dma_handle,
144                            0, idma->dma.device_id, idma);
145 }
146
147 static void iomd_free_dma(unsigned int chan, dma_t *dma)
148 {
149         struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma);
150
151         free_irq(idma->irq, idma);
152 }
153
154 static struct device isa_dma_dev = {
155         .init_name              = "fallback device",
156         .coherent_dma_mask      = ~(dma_addr_t)0,
157         .dma_mask               = &isa_dma_dev.coherent_dma_mask,
158 };
159
160 static void iomd_enable_dma(unsigned int chan, dma_t *dma)
161 {
162         struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma);
163         unsigned long dma_base = idma->base;
164         unsigned int ctrl = TRANSFER_SIZE | DMA_CR_E;
165
166         if (idma->dma.invalid) {
167                 idma->dma.invalid = 0;
168
169                 /*
170                  * Cope with ISA-style drivers which expect cache
171                  * coherence.
172                  */
173                 if (!idma->dma.sg) {
174                         idma->dma.sg = &idma->dma.buf;
175                         idma->dma.sgcount = 1;
176                         idma->dma.buf.length = idma->dma.count;
177                         idma->dma.buf.dma_address = dma_map_single(&isa_dma_dev,
178                                 idma->dma.addr, idma->dma.count,
179                                 idma->dma.dma_mode == DMA_MODE_READ ?
180                                 DMA_FROM_DEVICE : DMA_TO_DEVICE);
181                 }
182
183                 iomd_writeb(DMA_CR_C, dma_base + CR);
184                 idma->state = DMA_ST_AB;
185         }
186
187         if (idma->dma.dma_mode == DMA_MODE_READ)
188                 ctrl |= DMA_CR_D;
189
190         iomd_writeb(ctrl, dma_base + CR);
191         enable_irq(idma->irq);
192 }
193
194 static void iomd_disable_dma(unsigned int chan, dma_t *dma)
195 {
196         struct iomd_dma *idma = container_of(dma, struct iomd_dma, dma);
197         unsigned long dma_base = idma->base;
198         unsigned long flags;
199
200         local_irq_save(flags);
201         if (idma->state != ~DMA_ST_AB)
202                 disable_irq(idma->irq);
203         iomd_writeb(0, dma_base + CR);
204         local_irq_restore(flags);
205 }
206
207 static int iomd_set_dma_speed(unsigned int chan, dma_t *dma, int cycle)
208 {
209         int tcr, speed;
210
211         if (cycle < 188)
212                 speed = 3;
213         else if (cycle <= 250)
214                 speed = 2;
215         else if (cycle < 438)
216                 speed = 1;
217         else
218                 speed = 0;
219
220         tcr = iomd_readb(IOMD_DMATCR);
221         speed &= 3;
222
223         switch (chan) {
224         case DMA_0:
225                 tcr = (tcr & ~0x03) | speed;
226                 break;
227
228         case DMA_1:
229                 tcr = (tcr & ~0x0c) | (speed << 2);
230                 break;
231
232         case DMA_2:
233                 tcr = (tcr & ~0x30) | (speed << 4);
234                 break;
235
236         case DMA_3:
237                 tcr = (tcr & ~0xc0) | (speed << 6);
238                 break;
239
240         default:
241                 break;
242         }
243
244         iomd_writeb(tcr, IOMD_DMATCR);
245
246         return speed;
247 }
248
249 static struct dma_ops iomd_dma_ops = {
250         .type           = "IOMD",
251         .request        = iomd_request_dma,
252         .free           = iomd_free_dma,
253         .enable         = iomd_enable_dma,
254         .disable        = iomd_disable_dma,
255         .setspeed       = iomd_set_dma_speed,
256 };
257
258 static struct fiq_handler fh = {
259         .name   = "floppydma"
260 };
261
262 struct floppy_dma {
263         struct dma_struct       dma;
264         unsigned int            fiq;
265 };
266
267 static void floppy_enable_dma(unsigned int chan, dma_t *dma)
268 {
269         struct floppy_dma *fdma = container_of(dma, struct floppy_dma, dma);
270         void *fiqhandler_start;
271         unsigned int fiqhandler_length;
272         struct pt_regs regs;
273
274         if (fdma->dma.sg)
275                 BUG();
276
277         if (fdma->dma.dma_mode == DMA_MODE_READ) {
278                 extern unsigned char floppy_fiqin_start, floppy_fiqin_end;
279                 fiqhandler_start = &floppy_fiqin_start;
280                 fiqhandler_length = &floppy_fiqin_end - &floppy_fiqin_start;
281         } else {
282                 extern unsigned char floppy_fiqout_start, floppy_fiqout_end;
283                 fiqhandler_start = &floppy_fiqout_start;
284                 fiqhandler_length = &floppy_fiqout_end - &floppy_fiqout_start;
285         }
286
287         regs.ARM_r9  = fdma->dma.count;
288         regs.ARM_r10 = (unsigned long)fdma->dma.addr;
289         regs.ARM_fp  = (unsigned long)FLOPPYDMA_BASE;
290
291         if (claim_fiq(&fh)) {
292                 printk("floppydma: couldn't claim FIQ.\n");
293                 return;
294         }
295
296         set_fiq_handler(fiqhandler_start, fiqhandler_length);
297         set_fiq_regs(&regs);
298         enable_fiq(fdma->fiq);
299 }
300
301 static void floppy_disable_dma(unsigned int chan, dma_t *dma)
302 {
303         struct floppy_dma *fdma = container_of(dma, struct floppy_dma, dma);
304         disable_fiq(fdma->fiq);
305         release_fiq(&fh);
306 }
307
308 static int floppy_get_residue(unsigned int chan, dma_t *dma)
309 {
310         struct pt_regs regs;
311         get_fiq_regs(&regs);
312         return regs.ARM_r9;
313 }
314
315 static struct dma_ops floppy_dma_ops = {
316         .type           = "FIQDMA",
317         .enable         = floppy_enable_dma,
318         .disable        = floppy_disable_dma,
319         .residue        = floppy_get_residue,
320 };
321
322 /*
323  * This is virtual DMA - we don't need anything here.
324  */
325 static void sound_enable_disable_dma(unsigned int chan, dma_t *dma)
326 {
327 }
328
329 static struct dma_ops sound_dma_ops = {
330         .type           = "VIRTUAL",
331         .enable         = sound_enable_disable_dma,
332         .disable        = sound_enable_disable_dma,
333 };
334
335 static struct iomd_dma iomd_dma[6];
336
337 static struct floppy_dma floppy_dma = {
338         .dma            = {
339                 .d_ops  = &floppy_dma_ops,
340         },
341         .fiq            = FIQ_FLOPPYDATA,
342 };
343
344 static dma_t sound_dma = {
345         .d_ops          = &sound_dma_ops,
346 };
347
348 static int __init rpc_dma_init(void)
349 {
350         unsigned int i;
351         int ret;
352
353         iomd_writeb(0, IOMD_IO0CR);
354         iomd_writeb(0, IOMD_IO1CR);
355         iomd_writeb(0, IOMD_IO2CR);
356         iomd_writeb(0, IOMD_IO3CR);
357
358         iomd_writeb(0xa0, IOMD_DMATCR);
359
360         /*
361          * Setup DMA channels 2,3 to be for podules
362          * and channels 0,1 for internal devices
363          */
364         iomd_writeb(DMA_EXT_IO3|DMA_EXT_IO2, IOMD_DMAEXT);
365
366         iomd_dma[DMA_0].base    = IOMD_IO0CURA;
367         iomd_dma[DMA_0].irq     = IRQ_DMA0;
368         iomd_dma[DMA_1].base    = IOMD_IO1CURA;
369         iomd_dma[DMA_1].irq     = IRQ_DMA1;
370         iomd_dma[DMA_2].base    = IOMD_IO2CURA;
371         iomd_dma[DMA_2].irq     = IRQ_DMA2;
372         iomd_dma[DMA_3].base    = IOMD_IO3CURA;
373         iomd_dma[DMA_3].irq     = IRQ_DMA3;
374         iomd_dma[DMA_S0].base   = IOMD_SD0CURA;
375         iomd_dma[DMA_S0].irq    = IRQ_DMAS0;
376         iomd_dma[DMA_S1].base   = IOMD_SD1CURA;
377         iomd_dma[DMA_S1].irq    = IRQ_DMAS1;
378
379         for (i = DMA_0; i <= DMA_S1; i++) {
380                 iomd_dma[i].dma.d_ops = &iomd_dma_ops;
381
382                 ret = isa_dma_add(i, &iomd_dma[i].dma);
383                 if (ret)
384                         printk("IOMDDMA%u: unable to register: %d\n", i, ret);
385         }
386
387         ret = isa_dma_add(DMA_VIRTUAL_FLOPPY, &floppy_dma.dma);
388         if (ret)
389                 printk("IOMDFLOPPY: unable to register: %d\n", ret);
390         ret = isa_dma_add(DMA_VIRTUAL_SOUND, &sound_dma);
391         if (ret)
392                 printk("IOMDSOUND: unable to register: %d\n", ret);
393         return 0;
394 }
395 core_initcall(rpc_dma_init);