Index: /trunk/src/VBox/VMM/VMMR0/HMSVMR0.cpp
===================================================================
--- /trunk/src/VBox/VMM/VMMR0/HMSVMR0.cpp	(revision 55829)
+++ /trunk/src/VBox/VMM/VMMR0/HMSVMR0.cpp	(revision 55830)
@@ -196,5 +196,5 @@
 {
     /** The host's rflags/eflags. */
-    RTCCUINTREG     uEflags;
+    RTCCUINTREG     fEFlags;
 #if HC_ARCH_BITS == 32
     uint32_t        u32Alignment0;
@@ -340,5 +340,5 @@
 
     /* Paranoid: Disable interrupt as, in theory, interrupt handlers might mess with EFER. */
-    RTCCUINTREG uEflags = ASMIntDisableFlags();
+    RTCCUINTREG fEFlags = ASMIntDisableFlags();
 
     /*
@@ -357,5 +357,5 @@
         if (!pCpu->fIgnoreAMDVInUseError)
         {
-            ASMSetFlags(uEflags);
+            ASMSetFlags(fEFlags);
             return VERR_SVM_IN_USE;
         }
@@ -369,5 +369,5 @@
 
     /* Restore interrupts. */
-    ASMSetFlags(uEflags);
+    ASMSetFlags(fEFlags);
 
     /*
@@ -405,5 +405,5 @@
 
     /* Paranoid: Disable interrupts as, in theory, interrupt handlers might mess with EFER. */
-    RTCCUINTREG uEflags = ASMIntDisableFlags();
+    RTCCUINTREG fEFlags = ASMIntDisableFlags();
 
     /* Turn off AMD-V in the EFER MSR. */
@@ -415,5 +415,5 @@
 
     /* Restore interrupts. */
-    ASMSetFlags(uEflags);
+    ASMSetFlags(fEFlags);
 
     return VINF_SUCCESS;
@@ -3025,9 +3025,9 @@
      * executing guest code.
      */
-    pSvmTransient->uEflags = ASMIntDisableFlags();
+    pSvmTransient->fEFlags = ASMIntDisableFlags();
     if (   VM_FF_IS_PENDING(pVM, VM_FF_EMT_RENDEZVOUS | VM_FF_TM_VIRTUAL_SYNC)
         || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HM_TO_R3_MASK))
     {
-        ASMSetFlags(pSvmTransient->uEflags);
+        ASMSetFlags(pSvmTransient->fEFlags);
         VMMRZCallRing3Enable(pVCpu);
         STAM_COUNTER_INC(&pVCpu->hm.s.StatSwitchHmToR3FF);
@@ -3036,5 +3036,5 @@
     if (RTThreadPreemptIsPending(NIL_RTTHREAD))
     {
-        ASMSetFlags(pSvmTransient->uEflags);
+        ASMSetFlags(pSvmTransient->fEFlags);
         VMMRZCallRing3Enable(pVCpu);
         STAM_COUNTER_INC(&pVCpu->hm.s.StatPendingHostIrq);
@@ -3239,5 +3239,5 @@
 
     Assert(!(ASMGetFlags() & X86_EFL_IF));
-    ASMSetFlags(pSvmTransient->uEflags);                        /* Enable interrupts. */
+    ASMSetFlags(pSvmTransient->fEFlags);                        /* Enable interrupts. */
     VMMRZCallRing3Enable(pVCpu);                                /* It is now safe to do longjmps to ring-3!!! */
 
@@ -4737,5 +4737,5 @@
 
     pVCpu->hm.s.fLoadSaveGuestXcr0 = (pCtx->cr4 & X86_CR4_OSXSAVE) && pCtx->aXcr[0] != ASMGetXcr0();
-    Log4(("hmR0SvmExitXsetbv: New XCR0=%#RX64 fLoadSaveGuestXcr0=%d (cr4=%RX64) rcStrict=%Rrc\n", 
+    Log4(("hmR0SvmExitXsetbv: New XCR0=%#RX64 fLoadSaveGuestXcr0=%d (cr4=%RX64) rcStrict=%Rrc\n",
           pCtx->aXcr[0], pVCpu->hm.s.fLoadSaveGuestXcr0, pCtx->cr4, VBOXSTRICTRC_VAL(rcStrict)));
 
Index: /trunk/src/VBox/VMM/VMMR0/HMVMXR0.cpp
===================================================================
--- /trunk/src/VBox/VMM/VMMR0/HMVMXR0.cpp	(revision 55829)
+++ /trunk/src/VBox/VMM/VMMR0/HMVMXR0.cpp	(revision 55830)
@@ -214,5 +214,5 @@
 {
     /** The host's rflags/eflags. */
-    RTCCUINTREG     uEflags;
+    RTCCUINTREG     fEFlags;
 #if HC_ARCH_BITS == 32
     uint32_t        u32Alignment0;
@@ -775,5 +775,5 @@
 
     /* Paranoid: Disable interrupts as, in theory, interrupt handlers might mess with CR4. */
-    RTCCUINTREG uEflags = ASMIntDisableFlags();
+    RTCCUINTREG fEFlags = ASMIntDisableFlags();
 
     /* Enable the VMX bit in CR4 if necessary. */
@@ -792,5 +792,5 @@
 
     /* Restore interrupts. */
-    ASMSetFlags(uEflags);
+    ASMSetFlags(fEFlags);
     return rc;
 }
@@ -807,5 +807,5 @@
 
     /* Paranoid: Disable interrupts as, in theory, interrupts handlers might mess with CR4. */
-    RTCCUINTREG uEflags = ASMIntDisableFlags();
+    RTCCUINTREG fEFlags = ASMIntDisableFlags();
 
     /* If we're for some reason not in VMX root mode, then don't leave it. */
@@ -824,5 +824,5 @@
 
     /* Restore interrupts. */
-    ASMSetFlags(uEflags);
+    ASMSetFlags(fEFlags);
     return rc;
 }
@@ -5244,5 +5244,5 @@
     PHMGLOBALCPUINFO pCpu;
     RTHCPHYS        HCPhysCpuPage;
-    RTCCUINTREG     uOldEflags;
+    RTCCUINTREG     fOldEFlags;
 
     AssertReturn(pVM->hm.s.pfnHost32ToGuest64R0, VERR_HM_NO_32_TO_64_SWITCHER);
@@ -5260,5 +5260,5 @@
 
     /* Disable interrupts. */
-    uOldEflags = ASMIntDisableFlags();
+    fOldEFlags = ASMIntDisableFlags();
 
 #ifdef VBOX_WITH_VMMR0_DISABLE_LAPIC_NMI
@@ -5298,5 +5298,5 @@
     {
         SUPR0ChangeCR4(0, ~X86_CR4_VMXE);
-        ASMSetFlags(uOldEflags);
+        ASMSetFlags(fOldEFlags);
         pVM->hm.s.vmx.HCPhysVmxEnableError = HCPhysCpuPage;
         return rc2;
@@ -5306,5 +5306,5 @@
     AssertRC(rc2);
     Assert(!(ASMGetFlags() & X86_EFL_IF));
-    ASMSetFlags(uOldEflags);
+    ASMSetFlags(fOldEFlags);
     return rc;
 }
@@ -8651,5 +8651,5 @@
      * executing guest code.
      */
-    pVmxTransient->uEflags = ASMIntDisableFlags();
+    pVmxTransient->fEFlags = ASMIntDisableFlags();
     if (  (   VM_FF_IS_PENDING(pVM, VM_FF_EMT_RENDEZVOUS | VM_FF_TM_VIRTUAL_SYNC)
            || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HM_TO_R3_MASK))
@@ -8658,5 +8658,5 @@
     {
         hmR0VmxClearEventVmcs(pVCpu);
-        ASMSetFlags(pVmxTransient->uEflags);
+        ASMSetFlags(pVmxTransient->fEFlags);
         VMMRZCallRing3Enable(pVCpu);
         STAM_COUNTER_INC(&pVCpu->hm.s.StatSwitchHmToR3FF);
@@ -8667,5 +8667,5 @@
     {
         hmR0VmxClearEventVmcs(pVCpu);
-        ASMSetFlags(pVmxTransient->uEflags);
+        ASMSetFlags(pVmxTransient->fEFlags);
         VMMRZCallRing3Enable(pVCpu);
         STAM_COUNTER_INC(&pVCpu->hm.s.StatPendingHostIrq);
@@ -8883,5 +8883,5 @@
     hmR0VmxCheckHostEferMsr(pVCpu);                                   /* Verify that VMRUN/VMLAUNCH didn't modify host EFER. */
 #endif
-    ASMSetFlags(pVmxTransient->uEflags);                              /* Enable interrupts. */
+    ASMSetFlags(pVmxTransient->fEFlags);                              /* Enable interrupts. */
     VMMRZCallRing3Enable(pVCpu);                                      /* It is now safe to do longjmps to ring-3!!! */
 
