]> git.kernelconcepts.de Git - karo-tx-linux.git/blobdiff - arch/x86/kernel/paravirt_patch_64.c
Merge branch 'x86-fpu-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git...
[karo-tx-linux.git] / arch / x86 / kernel / paravirt_patch_64.c
index 915a4c0b217c0f115045d60ead02824fba940fd2..f4fcf26c9fcecefcd3b3c07db7cc85676e26aba6 100644 (file)
@@ -20,6 +20,7 @@ DEF_NATIVE(, mov64, "mov %rdi, %rax");
 
 #if defined(CONFIG_PARAVIRT_SPINLOCKS)
 DEF_NATIVE(pv_lock_ops, queued_spin_unlock, "movb $0, (%rdi)");
+DEF_NATIVE(pv_lock_ops, vcpu_is_preempted, "xor %rax, %rax");
 #endif
 
 unsigned paravirt_patch_ident_32(void *insnbuf, unsigned len)
@@ -35,6 +36,7 @@ unsigned paravirt_patch_ident_64(void *insnbuf, unsigned len)
 }
 
 extern bool pv_is_native_spin_unlock(void);
+extern bool pv_is_native_vcpu_is_preempted(void);
 
 unsigned native_patch(u8 type, u16 clobbers, void *ibuf,
                      unsigned long addr, unsigned len)
@@ -66,9 +68,19 @@ unsigned native_patch(u8 type, u16 clobbers, void *ibuf,
                                end   = end_pv_lock_ops_queued_spin_unlock;
                                goto patch_site;
                        }
+                       goto patch_default;
+
+               case PARAVIRT_PATCH(pv_lock_ops.vcpu_is_preempted):
+                       if (pv_is_native_vcpu_is_preempted()) {
+                               start = start_pv_lock_ops_vcpu_is_preempted;
+                               end   = end_pv_lock_ops_vcpu_is_preempted;
+                               goto patch_site;
+                       }
+                       goto patch_default;
 #endif
 
        default:
+patch_default:
                ret = paravirt_patch_default(type, clobbers, ibuf, addr, len);
                break;