2 * Copyright (C) 2003, Axis Communications AB.
7 #include <linux/interrupt.h>
9 #include <linux/kernel.h>
10 #include <linux/errno.h>
11 #include <linux/init.h>
12 #include <linux/profile.h>
14 #include <linux/of_irq.h>
15 #include <linux/proc_fs.h>
16 #include <linux/seq_file.h>
17 #include <linux/threads.h>
18 #include <linux/spinlock.h>
19 #include <linux/kernel_stat.h>
20 #include <hwregs/reg_map.h>
21 #include <hwregs/reg_rdwr.h>
22 #include <hwregs/intr_vect.h>
23 #include <hwregs/intr_vect_defs.h>
27 /* IRQ masks (refer to comment for crisv32_do_multiple) */
28 #if TIMER0_INTR_VECT - FIRST_IRQ < 32
29 #define TIMER_MASK (1 << (TIMER0_INTR_VECT - FIRST_IRQ))
32 #define TIMER_MASK (1 << (TIMER0_INTR_VECT - FIRST_IRQ - 32))
35 #ifdef CONFIG_ETRAX_KGDB
36 #if defined(CONFIG_ETRAX_KGDB_PORT0)
37 #define IGNOREMASK (1 << (SER0_INTR_VECT - FIRST_IRQ))
38 #elif defined(CONFIG_ETRAX_KGDB_PORT1)
39 #define IGNOREMASK (1 << (SER1_INTR_VECT - FIRST_IRQ))
40 #elif defined(CONFIG_ETRAX_KGB_PORT2)
41 #define IGNOREMASK (1 << (SER2_INTR_VECT - FIRST_IRQ))
42 #elif defined(CONFIG_ETRAX_KGDB_PORT3)
43 #define IGNOREMASK (1 << (SER3_INTR_VECT - FIRST_IRQ))
47 DEFINE_SPINLOCK(irq_lock);
49 struct cris_irq_allocation
51 int cpu; /* The CPU to which the IRQ is currently allocated. */
52 cpumask_t mask; /* The CPUs to which the IRQ may be allocated. */
55 struct cris_irq_allocation irq_allocations[NR_REAL_IRQS] =
56 { [0 ... NR_REAL_IRQS - 1] = {0, CPU_MASK_ALL} };
58 static unsigned long irq_regs[NR_CPUS] =
72 unsigned long cpu_irq_counters[NR_CPUS];
73 unsigned long irq_counters[NR_REAL_IRQS];
76 extern void weird_irq(void);
79 extern void system_call(void);
80 extern void nmi_interrupt(void);
81 extern void multiple_interrupt(void);
82 extern void gdb_handle_exception(void);
83 extern void i_mmu_refill(void);
84 extern void i_mmu_invalid(void);
85 extern void i_mmu_access(void);
86 extern void i_mmu_execute(void);
87 extern void d_mmu_refill(void);
88 extern void d_mmu_invalid(void);
89 extern void d_mmu_access(void);
90 extern void d_mmu_write(void);
93 extern void kgdb_init(void);
94 extern void breakpoint(void);
97 extern void breakh_BUG(void);
100 * Build the IRQ handler stubs using macros from irq.h.
102 #ifdef CONFIG_CRIS_MACH_ARTPEC3
103 BUILD_TIMER_IRQ(0x31, 0)
132 #ifdef CONFIG_ETRAXFS
133 BUILD_TIMER_IRQ(0x4b, 0)
177 /* Pointers to the low-level handlers. */
178 static void (*interrupt[MACH_IRQS])(void) = {
179 IRQ0x31_interrupt, IRQ0x32_interrupt, IRQ0x33_interrupt,
180 IRQ0x34_interrupt, IRQ0x35_interrupt, IRQ0x36_interrupt,
181 IRQ0x37_interrupt, IRQ0x38_interrupt, IRQ0x39_interrupt,
182 IRQ0x3a_interrupt, IRQ0x3b_interrupt, IRQ0x3c_interrupt,
183 IRQ0x3d_interrupt, IRQ0x3e_interrupt, IRQ0x3f_interrupt,
184 IRQ0x40_interrupt, IRQ0x41_interrupt, IRQ0x42_interrupt,
185 IRQ0x43_interrupt, IRQ0x44_interrupt, IRQ0x45_interrupt,
186 IRQ0x46_interrupt, IRQ0x47_interrupt, IRQ0x48_interrupt,
187 IRQ0x49_interrupt, IRQ0x4a_interrupt, IRQ0x4b_interrupt,
188 IRQ0x4c_interrupt, IRQ0x4d_interrupt, IRQ0x4e_interrupt,
189 IRQ0x4f_interrupt, IRQ0x50_interrupt,
191 IRQ0x51_interrupt, IRQ0x52_interrupt, IRQ0x53_interrupt,
192 IRQ0x54_interrupt, IRQ0x55_interrupt, IRQ0x56_interrupt,
193 IRQ0x57_interrupt, IRQ0x58_interrupt, IRQ0x59_interrupt,
194 IRQ0x5a_interrupt, IRQ0x5b_interrupt, IRQ0x5c_interrupt,
195 IRQ0x5d_interrupt, IRQ0x5e_interrupt, IRQ0x5f_interrupt,
196 IRQ0x60_interrupt, IRQ0x61_interrupt, IRQ0x62_interrupt,
197 IRQ0x63_interrupt, IRQ0x64_interrupt, IRQ0x65_interrupt,
198 IRQ0x66_interrupt, IRQ0x67_interrupt, IRQ0x68_interrupt,
199 IRQ0x69_interrupt, IRQ0x6a_interrupt, IRQ0x6b_interrupt,
200 IRQ0x6c_interrupt, IRQ0x6d_interrupt, IRQ0x6e_interrupt,
201 IRQ0x6f_interrupt, IRQ0x70_interrupt,
206 block_irq(int irq, int cpu)
211 spin_lock_irqsave(&irq_lock, flags);
212 /* Remember, 1 let thru, 0 block. */
213 if (irq - FIRST_IRQ < 32) {
214 intr_mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu],
216 intr_mask &= ~(1 << (irq - FIRST_IRQ));
217 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask,
220 intr_mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu],
222 intr_mask &= ~(1 << (irq - FIRST_IRQ - 32));
223 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask,
226 spin_unlock_irqrestore(&irq_lock, flags);
230 unblock_irq(int irq, int cpu)
235 spin_lock_irqsave(&irq_lock, flags);
236 /* Remember, 1 let thru, 0 block. */
237 if (irq - FIRST_IRQ < 32) {
238 intr_mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu],
240 intr_mask |= (1 << (irq - FIRST_IRQ));
241 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask,
244 intr_mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu],
246 intr_mask |= (1 << (irq - FIRST_IRQ - 32));
247 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask,
250 spin_unlock_irqrestore(&irq_lock, flags);
253 /* Find out which CPU the irq should be allocated to. */
254 static int irq_cpu(int irq)
259 spin_lock_irqsave(&irq_lock, flags);
260 cpu = irq_allocations[irq - FIRST_IRQ].cpu;
262 /* Fixed interrupts stay on the local CPU. */
263 if (cpu == CPU_FIXED)
265 spin_unlock_irqrestore(&irq_lock, flags);
266 return smp_processor_id();
270 /* Let the interrupt stay if possible */
271 if (cpumask_test_cpu(cpu, &irq_allocations[irq - FIRST_IRQ].mask))
274 /* IRQ must be moved to another CPU. */
275 cpu = cpumask_first(&irq_allocations[irq - FIRST_IRQ].mask);
276 irq_allocations[irq - FIRST_IRQ].cpu = cpu;
278 spin_unlock_irqrestore(&irq_lock, flags);
282 void crisv32_mask_irq(int irq)
286 for (cpu = 0; cpu < NR_CPUS; cpu++)
290 void crisv32_unmask_irq(int irq)
292 unblock_irq(irq, irq_cpu(irq));
296 static void enable_crisv32_irq(struct irq_data *data)
298 crisv32_unmask_irq(data->irq);
301 static void disable_crisv32_irq(struct irq_data *data)
303 crisv32_mask_irq(data->irq);
306 static int set_affinity_crisv32_irq(struct irq_data *data,
307 const struct cpumask *dest, bool force)
311 spin_lock_irqsave(&irq_lock, flags);
312 irq_allocations[data->irq - FIRST_IRQ].mask = *dest;
313 spin_unlock_irqrestore(&irq_lock, flags);
317 static struct irq_chip crisv32_irq_type = {
319 .irq_shutdown = disable_crisv32_irq,
320 .irq_enable = enable_crisv32_irq,
321 .irq_disable = disable_crisv32_irq,
322 .irq_set_affinity = set_affinity_crisv32_irq,
326 set_exception_vector(int n, irqvectptr addr)
328 etrax_irv->v[n] = (irqvectptr) addr;
331 extern void do_IRQ(int irq, struct pt_regs * regs);
334 crisv32_do_IRQ(int irq, int block, struct pt_regs* regs)
336 /* Interrupts that may not be moved to another CPU may
337 * skip blocking. This is currently only valid for the
338 * timer IRQ and the IPI and is used for the timer
339 * interrupt to avoid watchdog starvation.
346 block_irq(irq, smp_processor_id());
349 unblock_irq(irq, irq_cpu(irq));
352 /* If multiple interrupts occur simultaneously we get a multiple
353 * interrupt from the CPU and software has to sort out which
354 * interrupts that happened. There are two special cases here:
356 * 1. Timer interrupts may never be blocked because of the
357 * watchdog (refer to comment in include/asr/arch/irq.h)
358 * 2. GDB serial port IRQs are unhandled here and will be handled
359 * as a single IRQ when it strikes again because the GDB
360 * stubb wants to save the registers in its own fashion.
363 crisv32_do_multiple(struct pt_regs* regs)
367 int masked[NBR_REGS];
371 cpu = smp_processor_id();
373 /* An extra irq_enter here to prevent softIRQs to run after
374 * each do_IRQ. This will decrease the interrupt latency.
378 for (i = 0; i < NBR_REGS; i++) {
379 /* Get which IRQs that happened. */
380 masked[i] = REG_RD_INT_VECT(intr_vect, irq_regs[cpu],
383 /* Calculate new IRQ mask with these IRQs disabled. */
384 mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu], rw_mask, i);
387 /* Timer IRQ is never masked */
389 if ((i == 1) && (masked[0] & TIMER_MASK))
392 if ((i == 0) && (masked[0] & TIMER_MASK))
395 /* Block all the IRQs */
396 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask, i, mask);
398 /* Check for timer IRQ and handle it special. */
400 if ((i == 1) && (masked[i] & TIMER_MASK)) {
401 masked[i] &= ~TIMER_MASK;
402 do_IRQ(TIMER0_INTR_VECT, regs);
405 if ((i == 0) && (masked[i] & TIMER_MASK)) {
406 masked[i] &= ~TIMER_MASK;
407 do_IRQ(TIMER0_INTR_VECT, regs);
413 /* Remove IRQs that can't be handled as multiple. */
414 masked[0] &= ~IGNORE_MASK;
417 /* Handle the rest of the IRQs. */
418 for (i = 0; i < NBR_REGS; i++) {
419 for (bit = 0; bit < 32; bit++) {
420 if (masked[i] & (1 << bit))
421 do_IRQ(bit + FIRST_IRQ + i*32, regs);
425 /* Unblock all the IRQs. */
426 for (i = 0; i < NBR_REGS; i++) {
427 mask = REG_RD_INT_VECT(intr_vect, irq_regs[cpu], rw_mask, i);
429 REG_WR_INT_VECT(intr_vect, irq_regs[cpu], rw_mask, i, mask);
432 /* This irq_exit() will trigger the soft IRQs. */
436 static int crisv32_irq_map(struct irq_domain *h, unsigned int virq,
437 irq_hw_number_t hw_irq_num)
439 irq_set_chip_and_handler(virq, &crisv32_irq_type, handle_simple_irq);
444 static struct irq_domain_ops crisv32_irq_ops = {
445 .map = crisv32_irq_map,
446 .xlate = irq_domain_xlate_onecell,
450 * This is called by start_kernel. It fixes the IRQ masks and setup the
451 * interrupt vector table to point to bad_interrupt pointers.
458 reg_intr_vect_rw_mask vect_mask = {0};
459 struct device_node *np;
460 struct irq_domain *domain;
462 /* Clear all interrupts masks. */
463 for (i = 0; i < NBR_REGS; i++)
464 REG_WR_VECT(intr_vect, regi_irq, rw_mask, i, vect_mask);
466 for (i = 0; i < 256; i++)
467 etrax_irv->v[i] = weird_irq;
469 np = of_find_compatible_node(NULL, NULL, "axis,crisv32-intc");
470 domain = irq_domain_add_legacy(np, NR_IRQS - FIRST_IRQ,
471 FIRST_IRQ, FIRST_IRQ,
472 &crisv32_irq_ops, NULL);
474 irq_set_default_host(domain);
477 for (i = FIRST_IRQ, j = 0; j < NR_IRQS; i++, j++) {
478 set_exception_vector(i, interrupt[j]);
481 /* Mark Timer and IPI IRQs as CPU local */
482 irq_allocations[TIMER0_INTR_VECT - FIRST_IRQ].cpu = CPU_FIXED;
483 irq_set_status_flags(TIMER0_INTR_VECT, IRQ_PER_CPU);
484 irq_allocations[IPI_INTR_VECT - FIRST_IRQ].cpu = CPU_FIXED;
485 irq_set_status_flags(IPI_INTR_VECT, IRQ_PER_CPU);
487 set_exception_vector(0x00, nmi_interrupt);
488 set_exception_vector(0x30, multiple_interrupt);
490 /* Set up handler for various MMU bus faults. */
491 set_exception_vector(0x04, i_mmu_refill);
492 set_exception_vector(0x05, i_mmu_invalid);
493 set_exception_vector(0x06, i_mmu_access);
494 set_exception_vector(0x07, i_mmu_execute);
495 set_exception_vector(0x08, d_mmu_refill);
496 set_exception_vector(0x09, d_mmu_invalid);
497 set_exception_vector(0x0a, d_mmu_access);
498 set_exception_vector(0x0b, d_mmu_write);
501 /* Break 14 handler, used to implement cheap BUG(). */
502 set_exception_vector(0x1e, breakh_BUG);
505 /* The system-call trap is reached by "break 13". */
506 set_exception_vector(0x1d, system_call);
508 /* Exception handlers for debugging, both user-mode and kernel-mode. */
511 set_exception_vector(0x18, gdb_handle_exception);
512 /* Hardware single step. */
513 set_exception_vector(0x3, gdb_handle_exception);
514 /* Hardware breakpoint. */
515 set_exception_vector(0xc, gdb_handle_exception);
517 #ifdef CONFIG_ETRAX_KGDB
519 /* Everything is set up; now trap the kernel. */