| .. | .. |
|---|
| 2 | 2 | #ifndef _LINUX_IRQ_WORK_H |
|---|
| 3 | 3 | #define _LINUX_IRQ_WORK_H |
|---|
| 4 | 4 | |
|---|
| 5 | | -#include <linux/llist.h> |
|---|
| 5 | +#include <linux/smp_types.h> |
|---|
| 6 | 6 | |
|---|
| 7 | 7 | /* |
|---|
| 8 | 8 | * An entry can be in one of four states: |
|---|
| .. | .. |
|---|
| 13 | 13 | * busy NULL, 2 -> {free, claimed} : callback in progress, can be claimed |
|---|
| 14 | 14 | */ |
|---|
| 15 | 15 | |
|---|
| 16 | | -#define IRQ_WORK_PENDING BIT(0) |
|---|
| 17 | | -#define IRQ_WORK_BUSY BIT(1) |
|---|
| 18 | | - |
|---|
| 19 | | -/* Doesn't want IPI, wait for tick: */ |
|---|
| 20 | | -#define IRQ_WORK_LAZY BIT(2) |
|---|
| 21 | | -/* Run hard IRQ context, even on RT */ |
|---|
| 22 | | -#define IRQ_WORK_HARD_IRQ BIT(3) |
|---|
| 23 | | - |
|---|
| 24 | | -#define IRQ_WORK_CLAIMED (IRQ_WORK_PENDING | IRQ_WORK_BUSY) |
|---|
| 25 | | - |
|---|
| 26 | 16 | struct irq_work { |
|---|
| 27 | | - unsigned long flags; |
|---|
| 28 | | - struct llist_node llnode; |
|---|
| 17 | + union { |
|---|
| 18 | + struct __call_single_node node; |
|---|
| 19 | + struct { |
|---|
| 20 | + struct llist_node llnode; |
|---|
| 21 | + atomic_t flags; |
|---|
| 22 | + }; |
|---|
| 23 | + }; |
|---|
| 29 | 24 | void (*func)(struct irq_work *); |
|---|
| 30 | 25 | }; |
|---|
| 31 | 26 | |
|---|
| 32 | 27 | static inline |
|---|
| 33 | 28 | void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) |
|---|
| 34 | 29 | { |
|---|
| 35 | | - work->flags = 0; |
|---|
| 30 | + atomic_set(&work->flags, 0); |
|---|
| 36 | 31 | work->func = func; |
|---|
| 37 | 32 | } |
|---|
| 38 | 33 | |
|---|
| 39 | | -#define DEFINE_IRQ_WORK(name, _f) struct irq_work name = { .func = (_f), } |
|---|
| 34 | +#define DEFINE_IRQ_WORK(name, _f) struct irq_work name = { \ |
|---|
| 35 | + .flags = ATOMIC_INIT(0), \ |
|---|
| 36 | + .func = (_f) \ |
|---|
| 37 | +} |
|---|
| 38 | + |
|---|
| 40 | 39 | |
|---|
| 41 | 40 | bool irq_work_queue(struct irq_work *work); |
|---|
| 42 | 41 | bool irq_work_queue_on(struct irq_work *work, int cpu); |
|---|
| .. | .. |
|---|
| 49 | 48 | |
|---|
| 50 | 49 | void irq_work_run(void); |
|---|
| 51 | 50 | bool irq_work_needs_cpu(void); |
|---|
| 51 | +void irq_work_single(void *arg); |
|---|
| 52 | 52 | #else |
|---|
| 53 | 53 | static inline bool irq_work_needs_cpu(void) { return false; } |
|---|
| 54 | 54 | static inline void irq_work_run(void) { } |
|---|
| 55 | | -#endif |
|---|
| 56 | | - |
|---|
| 57 | | -#if defined(CONFIG_IRQ_WORK) && defined(CONFIG_PREEMPT_RT_FULL) |
|---|
| 58 | | -void irq_work_tick_soft(void); |
|---|
| 59 | | -#else |
|---|
| 60 | | -static inline void irq_work_tick_soft(void) { } |
|---|
| 55 | +static inline void irq_work_single(void *arg) { } |
|---|
| 61 | 56 | #endif |
|---|
| 62 | 57 | |
|---|
| 63 | 58 | #endif /* _LINUX_IRQ_WORK_H */ |
|---|