blob: 370dc9fa0b11916b97bc918d6f639d380bbbd62d [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/* dma.c: DMA controller management on FR401 and the like
2 *
3 * Copyright (C) 2004 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#include <linux/module.h>
13#include <linux/sched.h>
14#include <linux/spinlock.h>
15#include <linux/errno.h>
16#include <linux/init.h>
17#include <asm/dma.h>
18#include <asm/gpio-regs.h>
19#include <asm/irc-regs.h>
20#include <asm/cpu-irqs.h>
21
22struct frv_dma_channel {
23 uint8_t flags;
24#define FRV_DMA_FLAGS_RESERVED 0x01
25#define FRV_DMA_FLAGS_INUSE 0x02
26#define FRV_DMA_FLAGS_PAUSED 0x04
27 uint8_t cap; /* capabilities available */
28 int irq; /* completion IRQ */
29 uint32_t dreqbit;
30 uint32_t dackbit;
31 uint32_t donebit;
32 const unsigned long ioaddr; /* DMA controller regs addr */
33 const char *devname;
34 dma_irq_handler_t handler;
35 void *data;
36};
37
38
39#define __get_DMAC(IO,X) ({ *(volatile unsigned long *)((IO) + DMAC_##X##x); })
40
41#define __set_DMAC(IO,X,V) \
42do { \
43 *(volatile unsigned long *)((IO) + DMAC_##X##x) = (V); \
44 mb(); \
45} while(0)
46
47#define ___set_DMAC(IO,X,V) \
48do { \
49 *(volatile unsigned long *)((IO) + DMAC_##X##x) = (V); \
50} while(0)
51
52
53static struct frv_dma_channel frv_dma_channels[FRV_DMA_NCHANS] = {
54 [0] = {
55 .cap = FRV_DMA_CAP_DREQ | FRV_DMA_CAP_DACK | FRV_DMA_CAP_DONE,
56 .irq = IRQ_CPU_DMA0,
57 .dreqbit = SIR_DREQ0_INPUT,
58 .dackbit = SOR_DACK0_OUTPUT,
59 .donebit = SOR_DONE0_OUTPUT,
60 .ioaddr = 0xfe000900,
61 },
62 [1] = {
63 .cap = FRV_DMA_CAP_DREQ | FRV_DMA_CAP_DACK | FRV_DMA_CAP_DONE,
64 .irq = IRQ_CPU_DMA1,
65 .dreqbit = SIR_DREQ1_INPUT,
66 .dackbit = SOR_DACK1_OUTPUT,
67 .donebit = SOR_DONE1_OUTPUT,
68 .ioaddr = 0xfe000980,
69 },
70 [2] = {
71 .cap = FRV_DMA_CAP_DREQ | FRV_DMA_CAP_DACK,
72 .irq = IRQ_CPU_DMA2,
73 .dreqbit = SIR_DREQ2_INPUT,
74 .dackbit = SOR_DACK2_OUTPUT,
75 .ioaddr = 0xfe000a00,
76 },
77 [3] = {
78 .cap = FRV_DMA_CAP_DREQ | FRV_DMA_CAP_DACK,
79 .irq = IRQ_CPU_DMA3,
80 .dreqbit = SIR_DREQ3_INPUT,
81 .dackbit = SOR_DACK3_OUTPUT,
82 .ioaddr = 0xfe000a80,
83 },
84 [4] = {
85 .cap = FRV_DMA_CAP_DREQ,
86 .irq = IRQ_CPU_DMA4,
87 .dreqbit = SIR_DREQ4_INPUT,
88 .ioaddr = 0xfe001000,
89 },
90 [5] = {
91 .cap = FRV_DMA_CAP_DREQ,
92 .irq = IRQ_CPU_DMA5,
93 .dreqbit = SIR_DREQ5_INPUT,
94 .ioaddr = 0xfe001080,
95 },
96 [6] = {
97 .cap = FRV_DMA_CAP_DREQ,
98 .irq = IRQ_CPU_DMA6,
99 .dreqbit = SIR_DREQ6_INPUT,
100 .ioaddr = 0xfe001100,
101 },
102 [7] = {
103 .cap = FRV_DMA_CAP_DREQ,
104 .irq = IRQ_CPU_DMA7,
105 .dreqbit = SIR_DREQ7_INPUT,
106 .ioaddr = 0xfe001180,
107 },
108};
109
110static DEFINE_RWLOCK(frv_dma_channels_lock);
111
Peter Zijlstrab0d80032015-04-24 00:49:20 +0200112unsigned int frv_dma_inprogress;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700113
114#define frv_clear_dma_inprogress(channel) \
Peter Zijlstrab0d80032015-04-24 00:49:20 +0200115 (void)__atomic32_fetch_and(~(1 << (channel)), &frv_dma_inprogress);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700116
117#define frv_set_dma_inprogress(channel) \
Peter Zijlstrab0d80032015-04-24 00:49:20 +0200118 (void)__atomic32_fetch_or(1 << (channel), &frv_dma_inprogress);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700119
120/*****************************************************************************/
121/*
122 * DMA irq handler - determine channel involved, grab status and call real handler
123 */
David Howells7d12e782006-10-05 14:55:46 +0100124static irqreturn_t dma_irq_handler(int irq, void *_channel)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700125{
126 struct frv_dma_channel *channel = _channel;
127
128 frv_clear_dma_inprogress(channel - frv_dma_channels);
129 return channel->handler(channel - frv_dma_channels,
130 __get_DMAC(channel->ioaddr, CSTR),
David Howells7d12e782006-10-05 14:55:46 +0100131 channel->data);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700132
133} /* end dma_irq_handler() */
134
135/*****************************************************************************/
136/*
137 * Determine which DMA controllers are present on this CPU
138 */
139void __init frv_dma_init(void)
140{
141 unsigned long psr = __get_PSR();
142 int num_dma, i;
143
144 /* First, determine how many DMA channels are available */
145 switch (PSR_IMPLE(psr)) {
146 case PSR_IMPLE_FR405:
147 case PSR_IMPLE_FR451:
148 case PSR_IMPLE_FR501:
149 case PSR_IMPLE_FR551:
150 num_dma = FRV_DMA_8CHANS;
151 break;
152
153 case PSR_IMPLE_FR401:
154 default:
155 num_dma = FRV_DMA_4CHANS;
156 break;
157 }
158
159 /* Now mark all of the non-existent channels as reserved */
160 for(i = num_dma; i < FRV_DMA_NCHANS; i++)
161 frv_dma_channels[i].flags = FRV_DMA_FLAGS_RESERVED;
162
163} /* end frv_dma_init() */
164
165/*****************************************************************************/
166/*
167 * allocate a DMA controller channel and the IRQ associated with it
168 */
169int frv_dma_open(const char *devname,
170 unsigned long dmamask,
171 int dmacap,
172 dma_irq_handler_t handler,
173 unsigned long irq_flags,
174 void *data)
175{
176 struct frv_dma_channel *channel;
177 int dma, ret;
178 uint32_t val;
179
180 write_lock(&frv_dma_channels_lock);
181
182 ret = -ENOSPC;
183
184 for (dma = FRV_DMA_NCHANS - 1; dma >= 0; dma--) {
185 channel = &frv_dma_channels[dma];
186
187 if (!test_bit(dma, &dmamask))
188 continue;
189
190 if ((channel->cap & dmacap) != dmacap)
191 continue;
192
193 if (!frv_dma_channels[dma].flags)
194 goto found;
195 }
196
197 goto out;
198
199 found:
200 ret = request_irq(channel->irq, dma_irq_handler, irq_flags, devname, channel);
201 if (ret < 0)
202 goto out;
203
204 /* okay, we've allocated all the resources */
205 channel = &frv_dma_channels[dma];
206
207 channel->flags |= FRV_DMA_FLAGS_INUSE;
208 channel->devname = devname;
209 channel->handler = handler;
210 channel->data = data;
211
212 /* Now make sure we are set up for DMA and not GPIO */
213 /* SIR bit must be set for DMA to work */
214 __set_SIR(channel->dreqbit | __get_SIR());
215 /* SOR bits depend on what the caller requests */
216 val = __get_SOR();
217 if(dmacap & FRV_DMA_CAP_DACK)
218 val |= channel->dackbit;
219 else
220 val &= ~channel->dackbit;
221 if(dmacap & FRV_DMA_CAP_DONE)
222 val |= channel->donebit;
223 else
224 val &= ~channel->donebit;
225 __set_SOR(val);
226
227 ret = dma;
228 out:
229 write_unlock(&frv_dma_channels_lock);
230 return ret;
231} /* end frv_dma_open() */
232
233EXPORT_SYMBOL(frv_dma_open);
234
235/*****************************************************************************/
236/*
237 * close a DMA channel and its associated interrupt
238 */
239void frv_dma_close(int dma)
240{
241 struct frv_dma_channel *channel = &frv_dma_channels[dma];
242 unsigned long flags;
243
244 write_lock_irqsave(&frv_dma_channels_lock, flags);
245
246 free_irq(channel->irq, channel);
247 frv_dma_stop(dma);
248
249 channel->flags &= ~FRV_DMA_FLAGS_INUSE;
250
251 write_unlock_irqrestore(&frv_dma_channels_lock, flags);
252} /* end frv_dma_close() */
253
254EXPORT_SYMBOL(frv_dma_close);
255
256/*****************************************************************************/
257/*
258 * set static configuration on a DMA channel
259 */
260void frv_dma_config(int dma, unsigned long ccfr, unsigned long cctr, unsigned long apr)
261{
262 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
263
264 ___set_DMAC(ioaddr, CCFR, ccfr);
265 ___set_DMAC(ioaddr, CCTR, cctr);
266 ___set_DMAC(ioaddr, APR, apr);
267 mb();
268
269} /* end frv_dma_config() */
270
271EXPORT_SYMBOL(frv_dma_config);
272
273/*****************************************************************************/
274/*
275 * start a DMA channel
276 */
277void frv_dma_start(int dma,
278 unsigned long sba, unsigned long dba,
279 unsigned long pix, unsigned long six, unsigned long bcl)
280{
281 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
282
283 ___set_DMAC(ioaddr, SBA, sba);
284 ___set_DMAC(ioaddr, DBA, dba);
285 ___set_DMAC(ioaddr, PIX, pix);
286 ___set_DMAC(ioaddr, SIX, six);
287 ___set_DMAC(ioaddr, BCL, bcl);
288 ___set_DMAC(ioaddr, CSTR, 0);
289 mb();
290
291 __set_DMAC(ioaddr, CCTR, __get_DMAC(ioaddr, CCTR) | DMAC_CCTRx_ACT);
292 frv_set_dma_inprogress(dma);
293
294} /* end frv_dma_start() */
295
296EXPORT_SYMBOL(frv_dma_start);
297
298/*****************************************************************************/
299/*
300 * restart a DMA channel that's been stopped in circular addressing mode by comparison-end
301 */
302void frv_dma_restart_circular(int dma, unsigned long six)
303{
304 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
305
306 ___set_DMAC(ioaddr, SIX, six);
307 ___set_DMAC(ioaddr, CSTR, __get_DMAC(ioaddr, CSTR) & ~DMAC_CSTRx_CE);
308 mb();
309
310 __set_DMAC(ioaddr, CCTR, __get_DMAC(ioaddr, CCTR) | DMAC_CCTRx_ACT);
311 frv_set_dma_inprogress(dma);
312
313} /* end frv_dma_restart_circular() */
314
315EXPORT_SYMBOL(frv_dma_restart_circular);
316
317/*****************************************************************************/
318/*
319 * stop a DMA channel
320 */
321void frv_dma_stop(int dma)
322{
323 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
324 uint32_t cctr;
325
326 ___set_DMAC(ioaddr, CSTR, 0);
327 cctr = __get_DMAC(ioaddr, CCTR);
328 cctr &= ~(DMAC_CCTRx_IE | DMAC_CCTRx_ACT);
329 cctr |= DMAC_CCTRx_FC; /* fifo clear */
330 __set_DMAC(ioaddr, CCTR, cctr);
331 __set_DMAC(ioaddr, BCL, 0);
332 frv_clear_dma_inprogress(dma);
333} /* end frv_dma_stop() */
334
335EXPORT_SYMBOL(frv_dma_stop);
336
337/*****************************************************************************/
338/*
339 * test interrupt status of DMA channel
340 */
341int is_frv_dma_interrupting(int dma)
342{
343 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
344
345 return __get_DMAC(ioaddr, CSTR) & (1 << 23);
346
347} /* end is_frv_dma_interrupting() */
348
349EXPORT_SYMBOL(is_frv_dma_interrupting);
350
351/*****************************************************************************/
352/*
353 * dump data about a DMA channel
354 */
355void frv_dma_dump(int dma)
356{
357 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
358 unsigned long cstr, pix, six, bcl;
359
360 cstr = __get_DMAC(ioaddr, CSTR);
361 pix = __get_DMAC(ioaddr, PIX);
362 six = __get_DMAC(ioaddr, SIX);
363 bcl = __get_DMAC(ioaddr, BCL);
364
365 printk("DMA[%d] cstr=%lx pix=%lx six=%lx bcl=%lx\n", dma, cstr, pix, six, bcl);
366
367} /* end frv_dma_dump() */
368
369EXPORT_SYMBOL(frv_dma_dump);
370
371/*****************************************************************************/
372/*
373 * pause all DMA controllers
374 * - called by clock mangling routines
375 * - caller must be holding interrupts disabled
376 */
377void frv_dma_pause_all(void)
378{
379 struct frv_dma_channel *channel;
380 unsigned long ioaddr;
381 unsigned long cstr, cctr;
382 int dma;
383
384 write_lock(&frv_dma_channels_lock);
385
386 for (dma = FRV_DMA_NCHANS - 1; dma >= 0; dma--) {
387 channel = &frv_dma_channels[dma];
388
389 if (!(channel->flags & FRV_DMA_FLAGS_INUSE))
390 continue;
391
392 ioaddr = channel->ioaddr;
393 cctr = __get_DMAC(ioaddr, CCTR);
394 if (cctr & DMAC_CCTRx_ACT) {
395 cctr &= ~DMAC_CCTRx_ACT;
396 __set_DMAC(ioaddr, CCTR, cctr);
397
398 do {
399 cstr = __get_DMAC(ioaddr, CSTR);
400 } while (cstr & DMAC_CSTRx_BUSY);
401
402 if (cstr & DMAC_CSTRx_FED)
403 channel->flags |= FRV_DMA_FLAGS_PAUSED;
404 frv_clear_dma_inprogress(dma);
405 }
406 }
407
408} /* end frv_dma_pause_all() */
409
410EXPORT_SYMBOL(frv_dma_pause_all);
411
412/*****************************************************************************/
413/*
414 * resume paused DMA controllers
415 * - called by clock mangling routines
416 * - caller must be holding interrupts disabled
417 */
418void frv_dma_resume_all(void)
419{
420 struct frv_dma_channel *channel;
421 unsigned long ioaddr;
422 unsigned long cstr, cctr;
423 int dma;
424
425 for (dma = FRV_DMA_NCHANS - 1; dma >= 0; dma--) {
426 channel = &frv_dma_channels[dma];
427
428 if (!(channel->flags & FRV_DMA_FLAGS_PAUSED))
429 continue;
430
431 ioaddr = channel->ioaddr;
432 cstr = __get_DMAC(ioaddr, CSTR);
433 cstr &= ~(DMAC_CSTRx_FED | DMAC_CSTRx_INT);
434 __set_DMAC(ioaddr, CSTR, cstr);
435
436 cctr = __get_DMAC(ioaddr, CCTR);
437 cctr |= DMAC_CCTRx_ACT;
438 __set_DMAC(ioaddr, CCTR, cctr);
439
440 channel->flags &= ~FRV_DMA_FLAGS_PAUSED;
441 frv_set_dma_inprogress(dma);
442 }
443
444 write_unlock(&frv_dma_channels_lock);
445
446} /* end frv_dma_resume_all() */
447
448EXPORT_SYMBOL(frv_dma_resume_all);
449
450/*****************************************************************************/
451/*
452 * dma status clear
453 */
454void frv_dma_status_clear(int dma)
455{
456 unsigned long ioaddr = frv_dma_channels[dma].ioaddr;
457 uint32_t cctr;
458 ___set_DMAC(ioaddr, CSTR, 0);
459
460 cctr = __get_DMAC(ioaddr, CCTR);
461} /* end frv_dma_status_clear() */
462
463EXPORT_SYMBOL(frv_dma_status_clear);