summaryrefslogtreecommitdiff
path: root/arch/x86/include/asm/qspinlock.h
diff options
context:
space:
mode:
authorWaiman Long <longman@redhat.com>2017-02-20 13:36:03 -0500
committerPaolo Bonzini <pbonzini@redhat.com>2017-02-21 12:48:06 +0100
commit6c62985d576c8a816f528c39204207b9f449d923 (patch)
treeceeab09d6f75bcf723959a767d0456aff6ef78fa /arch/x86/include/asm/qspinlock.h
parent96794e4ed4d758272c486e1529e431efb7045265 (diff)
x86/paravirt: Change vcp_is_preempted() arg type to long
The cpu argument in the function prototype of vcpu_is_preempted() is changed from int to long. That makes it easier to provide a better optimized assembly version of that function. For Xen, vcpu_is_preempted(long) calls xen_vcpu_stolen(int), the downcast from long to int is not a problem as vCPU number won't exceed 32 bits. Signed-off-by: Waiman Long <longman@redhat.com> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Diffstat (limited to 'arch/x86/include/asm/qspinlock.h')
-rw-r--r--arch/x86/include/asm/qspinlock.h2
1 files changed, 1 insertions, 1 deletions
diff --git a/arch/x86/include/asm/qspinlock.h b/arch/x86/include/asm/qspinlock.h
index c343ab52579f..48a706f641f2 100644
--- a/arch/x86/include/asm/qspinlock.h
+++ b/arch/x86/include/asm/qspinlock.h
@@ -34,7 +34,7 @@ static inline void queued_spin_unlock(struct qspinlock *lock)
}
#define vcpu_is_preempted vcpu_is_preempted
-static inline bool vcpu_is_preempted(int cpu)
+static inline bool vcpu_is_preempted(long cpu)
{
return pv_vcpu_is_preempted(cpu);
}