such count is useful to assist decision make in cpuidle governor,
while w/o this patch only device interrupts through do_IRQ is
currently counted.
Signed-off-by: Kevin Tian <kevin.tian@intel.com>
struct cpu_user_regs *old_regs = set_irq_regs(regs);
ack_APIC_irq();
perfc_incr(apic_timer);
+ this_cpu(irq_count)++;
raise_softirq(TIMER_SOFTIRQ);
set_irq_regs(old_regs);
}
unsigned long v;
struct cpu_user_regs *old_regs = set_irq_regs(regs);
+ this_cpu(irq_count)++;
irq_enter();
/*
unsigned long v, v1;
struct cpu_user_regs *old_regs = set_irq_regs(regs);
+ this_cpu(irq_count)++;
irq_enter();
/* First tickle the hardware, only then report what went on. -- REW */
v = apic_read(APIC_ESR);
{
struct cpu_user_regs *old_regs = set_irq_regs(regs);
ack_APIC_irq();
+ this_cpu(irq_count)++;
hvm_do_pmu_interrupt(regs);
set_irq_regs(old_regs);
}
fastcall void smp_thermal_interrupt(struct cpu_user_regs *regs)
{
struct cpu_user_regs *old_regs = set_irq_regs(regs);
+ this_cpu(irq_count)++;
irq_enter();
vendor_thermal_interrupt(regs);
irq_exit();
struct cpu_user_regs *old_regs = set_irq_regs(regs);
ack_APIC_irq();
+ this_cpu(irq_count)++;
irq_enter();
mctc = mcheck_mca_logout(
struct cpu_user_regs *old_regs = set_irq_regs(regs);
ack_APIC_irq();
+ this_cpu(irq_count)++;
irq_enter();
me = smp_processor_id();
{
ack_APIC_irq();
perfc_incr(ipis);
+ this_cpu(irq_count)++;
irq_enter();
if ( !__sync_local_execstate() ||
(flush_flags & (FLUSH_TLB_GLOBAL | FLUSH_CACHE)) )
struct cpu_user_regs *old_regs = set_irq_regs(regs);
ack_APIC_irq();
perfc_incr(ipis);
+ this_cpu(irq_count)++;
set_irq_regs(old_regs);
}
ack_APIC_irq();
perfc_incr(ipis);
+ this_cpu(irq_count)++;
__smp_call_function_interrupt();
set_irq_regs(old_regs);
}