1/*
2 * linux/kernel/dma.c: A DMA channel allocator. Inspired by linux/kernel/irq.c.
3 *
4 * Written by Hennus Bergman, 1992.
5 *
6 * 1994/12/26: Changes by Alex Nash to fix a minor bug in /proc/dma.
7 *   In the previous version the reported device could end up being wrong,
8 *   if a device requested a DMA channel that was already in use.
9 *   [It also happened to remove the sizeof(char *) == sizeof(int)
10 *   assumption introduced because of those /proc/dma patches. -- Hennus]
11 */
12#include <linux/export.h>
13#include <linux/kernel.h>
14#include <linux/errno.h>
15#include <linux/spinlock.h>
16#include <linux/string.h>
17#include <linux/seq_file.h>
18#include <linux/proc_fs.h>
19#include <linux/init.h>
20#include <asm/dma.h>
21
22
23
24/* A note on resource allocation:
25 *
26 * All drivers needing DMA channels, should allocate and release them
27 * through the public routines `request_dma()' and `free_dma()'.
28 *
29 * In order to avoid problems, all processes should allocate resources in
30 * the same sequence and release them in the reverse order.
31 *
32 * So, when allocating DMAs and IRQs, first allocate the IRQ, then the DMA.
33 * When releasing them, first release the DMA, then release the IRQ.
34 * If you don't, you may cause allocation requests to fail unnecessarily.
35 * This doesn't really matter now, but it will once we get real semaphores
36 * in the kernel.
37 */
38
39
40DEFINE_SPINLOCK(dma_spin_lock);
41
42/*
43 *	If our port doesn't define this it has no PC like DMA
44 */
45
46#ifdef MAX_DMA_CHANNELS
47
48
49/* Channel n is busy iff dma_chan_busy[n].lock != 0.
50 * DMA0 used to be reserved for DRAM refresh, but apparently not any more...
51 * DMA4 is reserved for cascading.
52 */
53
54struct dma_chan {
55	int  lock;
56	const char *device_id;
57};
58
59static struct dma_chan dma_chan_busy[MAX_DMA_CHANNELS] = {
60	[4] = { 1, "cascade" },
61};
62
63
64/**
65 * request_dma - request and reserve a system DMA channel
66 * @dmanr: DMA channel number
67 * @device_id: reserving device ID string, used in /proc/dma
68 */
69int request_dma(unsigned int dmanr, const char * device_id)
70{
71	if (dmanr >= MAX_DMA_CHANNELS)
72		return -EINVAL;
73
74	if (xchg(&dma_chan_busy[dmanr].lock, 1) != 0)
75		return -EBUSY;
76
77	dma_chan_busy[dmanr].device_id = device_id;
78
79	/* old flag was 0, now contains 1 to indicate busy */
80	return 0;
81} /* request_dma */
82
83/**
84 * free_dma - free a reserved system DMA channel
85 * @dmanr: DMA channel number
86 */
87void free_dma(unsigned int dmanr)
88{
89	if (dmanr >= MAX_DMA_CHANNELS) {
90		printk(KERN_WARNING "Trying to free DMA%d\n", dmanr);
91		return;
92	}
93
94	if (xchg(&dma_chan_busy[dmanr].lock, 0) == 0) {
95		printk(KERN_WARNING "Trying to free free DMA%d\n", dmanr);
96		return;
97	}
98
99} /* free_dma */
100
101#else
102
103int request_dma(unsigned int dmanr, const char *device_id)
104{
105	return -EINVAL;
106}
107
108void free_dma(unsigned int dmanr)
109{
110}
111
112#endif
113
114#ifdef CONFIG_PROC_FS
115
116#ifdef MAX_DMA_CHANNELS
117static int proc_dma_show(struct seq_file *m, void *v)
118{
119	int i;
120
121	for (i = 0 ; i < MAX_DMA_CHANNELS ; i++) {
122		if (dma_chan_busy[i].lock) {
123			seq_printf(m, "%2d: %s\n", i,
124				   dma_chan_busy[i].device_id);
125		}
126	}
127	return 0;
128}
129#else
130static int proc_dma_show(struct seq_file *m, void *v)
131{
132	seq_puts(m, "No DMA\n");
133	return 0;
134}
135#endif /* MAX_DMA_CHANNELS */
136
137static int proc_dma_open(struct inode *inode, struct file *file)
138{
139	return single_open(file, proc_dma_show, NULL);
140}
141
142static const struct file_operations proc_dma_operations = {
143	.open		= proc_dma_open,
144	.read		= seq_read,
145	.llseek		= seq_lseek,
146	.release	= single_release,
147};
148
149static int __init proc_dma_init(void)
150{
151	proc_create("dma", 0, NULL, &proc_dma_operations);
152	return 0;
153}
154
155__initcall(proc_dma_init);
156#endif
157
158EXPORT_SYMBOL(request_dma);
159EXPORT_SYMBOL(free_dma);
160EXPORT_SYMBOL(dma_spin_lock);
161