Merge tag 'v3.3.7' into 3.3/master
[zen-stable.git] / arch / arm / mach-ixp4xx / ixp4xx_qmgr.c
blob852f7c9f87d06ff69c6defde901732a3926aea7d
1 /*
2 * Intel IXP4xx Queue Manager driver for Linux
4 * Copyright (C) 2007 Krzysztof Halasa <khc@pm.waw.pl>
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms of version 2 of the GNU General Public License
8 * as published by the Free Software Foundation.
9 */
11 #include <linux/ioport.h>
12 #include <linux/interrupt.h>
13 #include <linux/kernel.h>
14 #include <linux/module.h>
15 #include <mach/qmgr.h>
17 struct qmgr_regs __iomem *qmgr_regs;
18 static struct resource *mem_res;
19 static spinlock_t qmgr_lock;
20 static u32 used_sram_bitmap[4]; /* 128 16-dword pages */
21 static void (*irq_handlers[QUEUES])(void *pdev);
22 static void *irq_pdevs[QUEUES];
24 #if DEBUG_QMGR
25 char qmgr_queue_descs[QUEUES][32];
26 #endif
28 void qmgr_set_irq(unsigned int queue, int src,
29 void (*handler)(void *pdev), void *pdev)
31 unsigned long flags;
33 spin_lock_irqsave(&qmgr_lock, flags);
34 if (queue < HALF_QUEUES) {
35 u32 __iomem *reg;
36 int bit;
37 BUG_ON(src > QUEUE_IRQ_SRC_NOT_FULL);
38 reg = &qmgr_regs->irqsrc[queue >> 3]; /* 8 queues per u32 */
39 bit = (queue % 8) * 4; /* 3 bits + 1 reserved bit per queue */
40 __raw_writel((__raw_readl(reg) & ~(7 << bit)) | (src << bit),
41 reg);
42 } else
43 /* IRQ source for queues 32-63 is fixed */
44 BUG_ON(src != QUEUE_IRQ_SRC_NOT_NEARLY_EMPTY);
46 irq_handlers[queue] = handler;
47 irq_pdevs[queue] = pdev;
48 spin_unlock_irqrestore(&qmgr_lock, flags);
52 static irqreturn_t qmgr_irq1_a0(int irq, void *pdev)
54 int i, ret = 0;
55 u32 en_bitmap, src, stat;
57 /* ACK - it may clear any bits so don't rely on it */
58 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[0]);
60 en_bitmap = qmgr_regs->irqen[0];
61 while (en_bitmap) {
62 i = __fls(en_bitmap); /* number of the last "low" queue */
63 en_bitmap &= ~BIT(i);
64 src = qmgr_regs->irqsrc[i >> 3];
65 stat = qmgr_regs->stat1[i >> 3];
66 if (src & 4) /* the IRQ condition is inverted */
67 stat = ~stat;
68 if (stat & BIT(src & 3)) {
69 irq_handlers[i](irq_pdevs[i]);
70 ret = IRQ_HANDLED;
73 return ret;
77 static irqreturn_t qmgr_irq2_a0(int irq, void *pdev)
79 int i, ret = 0;
80 u32 req_bitmap;
82 /* ACK - it may clear any bits so don't rely on it */
83 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[1]);
85 req_bitmap = qmgr_regs->irqen[1] & qmgr_regs->statne_h;
86 while (req_bitmap) {
87 i = __fls(req_bitmap); /* number of the last "high" queue */
88 req_bitmap &= ~BIT(i);
89 irq_handlers[HALF_QUEUES + i](irq_pdevs[HALF_QUEUES + i]);
90 ret = IRQ_HANDLED;
92 return ret;
96 static irqreturn_t qmgr_irq(int irq, void *pdev)
98 int i, half = (irq == IRQ_IXP4XX_QM1 ? 0 : 1);
99 u32 req_bitmap = __raw_readl(&qmgr_regs->irqstat[half]);
101 if (!req_bitmap)
102 return 0;
103 __raw_writel(req_bitmap, &qmgr_regs->irqstat[half]); /* ACK */
105 while (req_bitmap) {
106 i = __fls(req_bitmap); /* number of the last queue */
107 req_bitmap &= ~BIT(i);
108 i += half * HALF_QUEUES;
109 irq_handlers[i](irq_pdevs[i]);
111 return IRQ_HANDLED;
115 void qmgr_enable_irq(unsigned int queue)
117 unsigned long flags;
118 int half = queue / 32;
119 u32 mask = 1 << (queue & (HALF_QUEUES - 1));
121 spin_lock_irqsave(&qmgr_lock, flags);
122 __raw_writel(__raw_readl(&qmgr_regs->irqen[half]) | mask,
123 &qmgr_regs->irqen[half]);
124 spin_unlock_irqrestore(&qmgr_lock, flags);
127 void qmgr_disable_irq(unsigned int queue)
129 unsigned long flags;
130 int half = queue / 32;
131 u32 mask = 1 << (queue & (HALF_QUEUES - 1));
133 spin_lock_irqsave(&qmgr_lock, flags);
134 __raw_writel(__raw_readl(&qmgr_regs->irqen[half]) & ~mask,
135 &qmgr_regs->irqen[half]);
136 __raw_writel(mask, &qmgr_regs->irqstat[half]); /* clear */
137 spin_unlock_irqrestore(&qmgr_lock, flags);
140 static inline void shift_mask(u32 *mask)
142 mask[3] = mask[3] << 1 | mask[2] >> 31;
143 mask[2] = mask[2] << 1 | mask[1] >> 31;
144 mask[1] = mask[1] << 1 | mask[0] >> 31;
145 mask[0] <<= 1;
148 #if DEBUG_QMGR
149 int qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
150 unsigned int nearly_empty_watermark,
151 unsigned int nearly_full_watermark,
152 const char *desc_format, const char* name)
153 #else
154 int __qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
155 unsigned int nearly_empty_watermark,
156 unsigned int nearly_full_watermark)
157 #endif
159 u32 cfg, addr = 0, mask[4]; /* in 16-dwords */
160 int err;
162 BUG_ON(queue >= QUEUES);
164 if ((nearly_empty_watermark | nearly_full_watermark) & ~7)
165 return -EINVAL;
167 switch (len) {
168 case 16:
169 cfg = 0 << 24;
170 mask[0] = 0x1;
171 break;
172 case 32:
173 cfg = 1 << 24;
174 mask[0] = 0x3;
175 break;
176 case 64:
177 cfg = 2 << 24;
178 mask[0] = 0xF;
179 break;
180 case 128:
181 cfg = 3 << 24;
182 mask[0] = 0xFF;
183 break;
184 default:
185 return -EINVAL;
188 cfg |= nearly_empty_watermark << 26;
189 cfg |= nearly_full_watermark << 29;
190 len /= 16; /* in 16-dwords: 1, 2, 4 or 8 */
191 mask[1] = mask[2] = mask[3] = 0;
193 if (!try_module_get(THIS_MODULE))
194 return -ENODEV;
196 spin_lock_irq(&qmgr_lock);
197 if (__raw_readl(&qmgr_regs->sram[queue])) {
198 err = -EBUSY;
199 goto err;
202 while (1) {
203 if (!(used_sram_bitmap[0] & mask[0]) &&
204 !(used_sram_bitmap[1] & mask[1]) &&
205 !(used_sram_bitmap[2] & mask[2]) &&
206 !(used_sram_bitmap[3] & mask[3]))
207 break; /* found free space */
209 addr++;
210 shift_mask(mask);
211 if (addr + len > ARRAY_SIZE(qmgr_regs->sram)) {
212 printk(KERN_ERR "qmgr: no free SRAM space for"
213 " queue %i\n", queue);
214 err = -ENOMEM;
215 goto err;
219 used_sram_bitmap[0] |= mask[0];
220 used_sram_bitmap[1] |= mask[1];
221 used_sram_bitmap[2] |= mask[2];
222 used_sram_bitmap[3] |= mask[3];
223 __raw_writel(cfg | (addr << 14), &qmgr_regs->sram[queue]);
224 #if DEBUG_QMGR
225 snprintf(qmgr_queue_descs[queue], sizeof(qmgr_queue_descs[0]),
226 desc_format, name);
227 printk(KERN_DEBUG "qmgr: requested queue %s(%i) addr = 0x%02X\n",
228 qmgr_queue_descs[queue], queue, addr);
229 #endif
230 spin_unlock_irq(&qmgr_lock);
231 return 0;
233 err:
234 spin_unlock_irq(&qmgr_lock);
235 module_put(THIS_MODULE);
236 return err;
239 void qmgr_release_queue(unsigned int queue)
241 u32 cfg, addr, mask[4];
243 BUG_ON(queue >= QUEUES); /* not in valid range */
245 spin_lock_irq(&qmgr_lock);
246 cfg = __raw_readl(&qmgr_regs->sram[queue]);
247 addr = (cfg >> 14) & 0xFF;
249 BUG_ON(!addr); /* not requested */
251 switch ((cfg >> 24) & 3) {
252 case 0: mask[0] = 0x1; break;
253 case 1: mask[0] = 0x3; break;
254 case 2: mask[0] = 0xF; break;
255 case 3: mask[0] = 0xFF; break;
258 mask[1] = mask[2] = mask[3] = 0;
260 while (addr--)
261 shift_mask(mask);
263 #if DEBUG_QMGR
264 printk(KERN_DEBUG "qmgr: releasing queue %s(%i)\n",
265 qmgr_queue_descs[queue], queue);
266 qmgr_queue_descs[queue][0] = '\x0';
267 #endif
269 while ((addr = qmgr_get_entry(queue)))
270 printk(KERN_ERR "qmgr: released queue %i not empty: 0x%08X\n",
271 queue, addr);
273 __raw_writel(0, &qmgr_regs->sram[queue]);
275 used_sram_bitmap[0] &= ~mask[0];
276 used_sram_bitmap[1] &= ~mask[1];
277 used_sram_bitmap[2] &= ~mask[2];
278 used_sram_bitmap[3] &= ~mask[3];
279 irq_handlers[queue] = NULL; /* catch IRQ bugs */
280 spin_unlock_irq(&qmgr_lock);
282 module_put(THIS_MODULE);
285 static int qmgr_init(void)
287 int i, err;
288 irq_handler_t handler1, handler2;
290 mem_res = request_mem_region(IXP4XX_QMGR_BASE_PHYS,
291 IXP4XX_QMGR_REGION_SIZE,
292 "IXP4xx Queue Manager");
293 if (mem_res == NULL)
294 return -EBUSY;
296 qmgr_regs = ioremap(IXP4XX_QMGR_BASE_PHYS, IXP4XX_QMGR_REGION_SIZE);
297 if (qmgr_regs == NULL) {
298 err = -ENOMEM;
299 goto error_map;
302 /* reset qmgr registers */
303 for (i = 0; i < 4; i++) {
304 __raw_writel(0x33333333, &qmgr_regs->stat1[i]);
305 __raw_writel(0, &qmgr_regs->irqsrc[i]);
307 for (i = 0; i < 2; i++) {
308 __raw_writel(0, &qmgr_regs->stat2[i]);
309 __raw_writel(0xFFFFFFFF, &qmgr_regs->irqstat[i]); /* clear */
310 __raw_writel(0, &qmgr_regs->irqen[i]);
313 __raw_writel(0xFFFFFFFF, &qmgr_regs->statne_h);
314 __raw_writel(0, &qmgr_regs->statf_h);
316 for (i = 0; i < QUEUES; i++)
317 __raw_writel(0, &qmgr_regs->sram[i]);
319 if (cpu_is_ixp42x_rev_a0()) {
320 handler1 = qmgr_irq1_a0;
321 handler2 = qmgr_irq2_a0;
322 } else
323 handler1 = handler2 = qmgr_irq;
325 err = request_irq(IRQ_IXP4XX_QM1, handler1, 0, "IXP4xx Queue Manager",
326 NULL);
327 if (err) {
328 printk(KERN_ERR "qmgr: failed to request IRQ%i (%i)\n",
329 IRQ_IXP4XX_QM1, err);
330 goto error_irq;
333 err = request_irq(IRQ_IXP4XX_QM2, handler2, 0, "IXP4xx Queue Manager",
334 NULL);
335 if (err) {
336 printk(KERN_ERR "qmgr: failed to request IRQ%i (%i)\n",
337 IRQ_IXP4XX_QM2, err);
338 goto error_irq2;
341 used_sram_bitmap[0] = 0xF; /* 4 first pages reserved for config */
342 spin_lock_init(&qmgr_lock);
344 printk(KERN_INFO "IXP4xx Queue Manager initialized.\n");
345 return 0;
347 error_irq2:
348 free_irq(IRQ_IXP4XX_QM1, NULL);
349 error_irq:
350 iounmap(qmgr_regs);
351 error_map:
352 release_mem_region(IXP4XX_QMGR_BASE_PHYS, IXP4XX_QMGR_REGION_SIZE);
353 return err;
356 static void qmgr_remove(void)
358 free_irq(IRQ_IXP4XX_QM1, NULL);
359 free_irq(IRQ_IXP4XX_QM2, NULL);
360 synchronize_irq(IRQ_IXP4XX_QM1);
361 synchronize_irq(IRQ_IXP4XX_QM2);
362 iounmap(qmgr_regs);
363 release_mem_region(IXP4XX_QMGR_BASE_PHYS, IXP4XX_QMGR_REGION_SIZE);
366 module_init(qmgr_init);
367 module_exit(qmgr_remove);
369 MODULE_LICENSE("GPL v2");
370 MODULE_AUTHOR("Krzysztof Halasa");
372 EXPORT_SYMBOL(qmgr_regs);
373 EXPORT_SYMBOL(qmgr_set_irq);
374 EXPORT_SYMBOL(qmgr_enable_irq);
375 EXPORT_SYMBOL(qmgr_disable_irq);
376 #if DEBUG_QMGR
377 EXPORT_SYMBOL(qmgr_queue_descs);
378 EXPORT_SYMBOL(qmgr_request_queue);
379 #else
380 EXPORT_SYMBOL(__qmgr_request_queue);
381 #endif
382 EXPORT_SYMBOL(qmgr_release_queue);