| .. | .. | 
|---|
| 34 | 34 |   * true and let the handler run. | 
|---|
| 35 | 35 |   */ | 
|---|
| 36 | 36 |  bool irq_wait_for_poll(struct irq_desc *desc) | 
|---|
 | 37 | +	__must_hold(&desc->lock)  | 
|---|
| 37 | 38 |  { | 
|---|
| 38 | 39 |  	if (WARN_ONCE(irq_poll_cpu == smp_processor_id(), | 
|---|
| 39 | 40 |  		      "irq poll in progress on cpu %d for irq %d\n", | 
|---|
| .. | .. | 
|---|
| 66 | 67 |  	raw_spin_lock(&desc->lock); | 
|---|
| 67 | 68 |   | 
|---|
| 68 | 69 |  	/* | 
|---|
| 69 |  | -	 * PER_CPU, nested thread interrupts and interrupts explicitely  | 
|---|
 | 70 | +	 * PER_CPU, nested thread interrupts and interrupts explicitly  | 
|---|
| 70 | 71 |  	 * marked polled are excluded from polling. | 
|---|
| 71 | 72 |  	 */ | 
|---|
| 72 | 73 |  	if (irq_settings_is_per_cpu(desc) || | 
|---|
| .. | .. | 
|---|
| 76 | 77 |   | 
|---|
| 77 | 78 |  	/* | 
|---|
| 78 | 79 |  	 * Do not poll disabled interrupts unless the spurious | 
|---|
| 79 |  | -	 * disabled poller asks explicitely.  | 
|---|
 | 80 | +	 * disabled poller asks explicitly.  | 
|---|
| 80 | 81 |  	 */ | 
|---|
| 81 | 82 |  	if (irqd_irq_disabled(&desc->irq_data) && !force) | 
|---|
| 82 | 83 |  		goto out; | 
|---|
| .. | .. | 
|---|
| 212 | 213 |  	 */ | 
|---|
| 213 | 214 |  	raw_spin_lock_irqsave(&desc->lock, flags); | 
|---|
| 214 | 215 |  	for_each_action_of_desc(desc, action) { | 
|---|
| 215 |  | -		printk(KERN_ERR "[<%p>] %pf", action->handler, action->handler);  | 
|---|
 | 216 | +		printk(KERN_ERR "[<%p>] %ps", action->handler, action->handler);  | 
|---|
| 216 | 217 |  		if (action->thread_fn) | 
|---|
| 217 |  | -			printk(KERN_CONT " threaded [<%p>] %pf",  | 
|---|
 | 218 | +			printk(KERN_CONT " threaded [<%p>] %ps",  | 
|---|
| 218 | 219 |  					action->thread_fn, action->thread_fn); | 
|---|
| 219 | 220 |  		printk(KERN_CONT "\n"); | 
|---|
| 220 | 221 |  	} | 
|---|
| .. | .. | 
|---|
| 292 | 293 |  	 * So in case a thread is woken, we just note the fact and | 
|---|
| 293 | 294 |  	 * defer the analysis to the next hardware interrupt. | 
|---|
| 294 | 295 |  	 * | 
|---|
| 295 |  | -	 * The threaded handlers store whether they sucessfully  | 
|---|
 | 296 | +	 * The threaded handlers store whether they successfully  | 
|---|
| 296 | 297 |  	 * handled an interrupt and we check whether that number | 
|---|
| 297 | 298 |  	 * changed versus the last invocation. | 
|---|
| 298 | 299 |  	 * | 
|---|
| .. | .. | 
|---|
| 442 | 443 |   | 
|---|
| 443 | 444 |  static int __init irqfixup_setup(char *str) | 
|---|
| 444 | 445 |  { | 
|---|
| 445 |  | -#ifdef CONFIG_PREEMPT_RT_BASE  | 
|---|
| 446 |  | -	pr_warn("irqfixup boot option not supported w/ CONFIG_PREEMPT_RT_BASE\n");  | 
|---|
 | 446 | +#ifdef CONFIG_PREEMPT_RT  | 
|---|
 | 447 | +	pr_warn("irqfixup boot option not supported w/ CONFIG_PREEMPT_RT\n");  | 
|---|
| 447 | 448 |  	return 1; | 
|---|
| 448 | 449 |  #endif | 
|---|
| 449 | 450 |  	irqfixup = 1; | 
|---|
| .. | .. | 
|---|
| 458 | 459 |   | 
|---|
| 459 | 460 |  static int __init irqpoll_setup(char *str) | 
|---|
| 460 | 461 |  { | 
|---|
| 461 |  | -#ifdef CONFIG_PREEMPT_RT_BASE  | 
|---|
| 462 |  | -	pr_warn("irqpoll boot option not supported w/ CONFIG_PREEMPT_RT_BASE\n");  | 
|---|
 | 462 | +#ifdef CONFIG_PREEMPT_RT  | 
|---|
 | 463 | +	pr_warn("irqpoll boot option not supported w/ CONFIG_PREEMPT_RT\n");  | 
|---|
| 463 | 464 |  	return 1; | 
|---|
| 464 | 465 |  #endif | 
|---|
| 465 | 466 |  	irqfixup = 2; | 
|---|