117 lines
2.3 KiB
C
117 lines
2.3 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2012 Regents of the University of California
|
|
* Copyright (C) 2017 SiFive
|
|
* Copyright (C) 2018 Christoph Hellwig
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/irqchip.h>
|
|
#include <linux/irqdomain.h>
|
|
#include <linux/module.h>
|
|
#include <linux/scs.h>
|
|
#include <linux/seq_file.h>
|
|
#include <asm/sbi.h>
|
|
#include <asm/smp.h>
|
|
#include <asm/softirq_stack.h>
|
|
#include <asm/stacktrace.h>
|
|
|
|
static struct fwnode_handle *(*__get_intc_node)(void);
|
|
|
|
void riscv_set_intc_hwnode_fn(struct fwnode_handle *(*fn)(void))
|
|
{
|
|
__get_intc_node = fn;
|
|
}
|
|
|
|
struct fwnode_handle *riscv_get_intc_hwnode(void)
|
|
{
|
|
if (__get_intc_node)
|
|
return __get_intc_node();
|
|
|
|
return NULL;
|
|
}
|
|
EXPORT_SYMBOL_GPL(riscv_get_intc_hwnode);
|
|
|
|
#ifdef CONFIG_IRQ_STACKS
|
|
#include <asm/irq_stack.h>
|
|
|
|
DECLARE_PER_CPU(ulong *, irq_shadow_call_stack_ptr);
|
|
|
|
#ifdef CONFIG_SHADOW_CALL_STACK
|
|
DEFINE_PER_CPU(ulong *, irq_shadow_call_stack_ptr);
|
|
#endif
|
|
|
|
static void init_irq_scs(void)
|
|
{
|
|
int cpu;
|
|
|
|
if (!scs_is_enabled())
|
|
return;
|
|
|
|
for_each_possible_cpu(cpu)
|
|
per_cpu(irq_shadow_call_stack_ptr, cpu) =
|
|
scs_alloc(cpu_to_node(cpu));
|
|
}
|
|
|
|
DEFINE_PER_CPU(ulong *, irq_stack_ptr);
|
|
|
|
#ifdef CONFIG_VMAP_STACK
|
|
static void init_irq_stacks(void)
|
|
{
|
|
int cpu;
|
|
ulong *p;
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
p = arch_alloc_vmap_stack(IRQ_STACK_SIZE, cpu_to_node(cpu));
|
|
per_cpu(irq_stack_ptr, cpu) = p;
|
|
}
|
|
}
|
|
#else
|
|
/* irq stack only needs to be 16 byte aligned - not IRQ_STACK_SIZE aligned. */
|
|
DEFINE_PER_CPU_ALIGNED(ulong [IRQ_STACK_SIZE/sizeof(ulong)], irq_stack);
|
|
|
|
static void init_irq_stacks(void)
|
|
{
|
|
int cpu;
|
|
|
|
for_each_possible_cpu(cpu)
|
|
per_cpu(irq_stack_ptr, cpu) = per_cpu(irq_stack, cpu);
|
|
}
|
|
#endif /* CONFIG_VMAP_STACK */
|
|
|
|
#ifdef CONFIG_SOFTIRQ_ON_OWN_STACK
|
|
static void ___do_softirq(struct pt_regs *regs)
|
|
{
|
|
__do_softirq();
|
|
}
|
|
|
|
void do_softirq_own_stack(void)
|
|
{
|
|
if (on_thread_stack())
|
|
call_on_irq_stack(NULL, ___do_softirq);
|
|
else
|
|
__do_softirq();
|
|
}
|
|
#endif /* CONFIG_SOFTIRQ_ON_OWN_STACK */
|
|
|
|
#else
|
|
static void init_irq_scs(void) {}
|
|
static void init_irq_stacks(void) {}
|
|
#endif /* CONFIG_IRQ_STACKS */
|
|
|
|
int arch_show_interrupts(struct seq_file *p, int prec)
|
|
{
|
|
show_ipi_stats(p, prec);
|
|
return 0;
|
|
}
|
|
|
|
void __init init_IRQ(void)
|
|
{
|
|
init_irq_scs();
|
|
init_irq_stacks();
|
|
irqchip_init();
|
|
if (!handle_arch_irq)
|
|
panic("No interrupt controller found.");
|
|
sbi_ipi_init();
|
|
}
|