.. | .. |
---|
2 | 2 | #ifndef _LINUX_IRQ_WORK_H |
---|
3 | 3 | #define _LINUX_IRQ_WORK_H |
---|
4 | 4 | |
---|
5 | | -#include <linux/llist.h> |
---|
| 5 | +#include <linux/smp_types.h> |
---|
6 | 6 | |
---|
7 | 7 | /* |
---|
8 | 8 | * An entry can be in one of four states: |
---|
.. | .. |
---|
13 | 13 | * busy NULL, 2 -> {free, claimed} : callback in progress, can be claimed |
---|
14 | 14 | */ |
---|
15 | 15 | |
---|
16 | | -#define IRQ_WORK_PENDING BIT(0) |
---|
17 | | -#define IRQ_WORK_BUSY BIT(1) |
---|
18 | | - |
---|
19 | | -/* Doesn't want IPI, wait for tick: */ |
---|
20 | | -#define IRQ_WORK_LAZY BIT(2) |
---|
21 | | -/* Run hard IRQ context, even on RT */ |
---|
22 | | -#define IRQ_WORK_HARD_IRQ BIT(3) |
---|
23 | | - |
---|
24 | | -#define IRQ_WORK_CLAIMED (IRQ_WORK_PENDING | IRQ_WORK_BUSY) |
---|
25 | | - |
---|
26 | 16 | struct irq_work { |
---|
27 | | - unsigned long flags; |
---|
28 | | - struct llist_node llnode; |
---|
| 17 | + union { |
---|
| 18 | + struct __call_single_node node; |
---|
| 19 | + struct { |
---|
| 20 | + struct llist_node llnode; |
---|
| 21 | + atomic_t flags; |
---|
| 22 | + }; |
---|
| 23 | + }; |
---|
29 | 24 | void (*func)(struct irq_work *); |
---|
30 | 25 | }; |
---|
31 | 26 | |
---|
32 | 27 | static inline |
---|
33 | 28 | void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) |
---|
34 | 29 | { |
---|
35 | | - work->flags = 0; |
---|
| 30 | + atomic_set(&work->flags, 0); |
---|
36 | 31 | work->func = func; |
---|
37 | 32 | } |
---|
38 | 33 | |
---|
39 | | -#define DEFINE_IRQ_WORK(name, _f) struct irq_work name = { .func = (_f), } |
---|
| 34 | +#define DEFINE_IRQ_WORK(name, _f) struct irq_work name = { \ |
---|
| 35 | + .flags = ATOMIC_INIT(0), \ |
---|
| 36 | + .func = (_f) \ |
---|
| 37 | +} |
---|
| 38 | + |
---|
40 | 39 | |
---|
41 | 40 | bool irq_work_queue(struct irq_work *work); |
---|
42 | 41 | bool irq_work_queue_on(struct irq_work *work, int cpu); |
---|
.. | .. |
---|
49 | 48 | |
---|
50 | 49 | void irq_work_run(void); |
---|
51 | 50 | bool irq_work_needs_cpu(void); |
---|
| 51 | +void irq_work_single(void *arg); |
---|
52 | 52 | #else |
---|
53 | 53 | static inline bool irq_work_needs_cpu(void) { return false; } |
---|
54 | 54 | static inline void irq_work_run(void) { } |
---|
55 | | -#endif |
---|
56 | | - |
---|
57 | | -#if defined(CONFIG_IRQ_WORK) && defined(CONFIG_PREEMPT_RT_FULL) |
---|
58 | | -void irq_work_tick_soft(void); |
---|
59 | | -#else |
---|
60 | | -static inline void irq_work_tick_soft(void) { } |
---|
| 55 | +static inline void irq_work_single(void *arg) { } |
---|
61 | 56 | #endif |
---|
62 | 57 | |
---|
63 | 58 | #endif /* _LINUX_IRQ_WORK_H */ |
---|