x86/retpoline: Optimize inline assembler for vmexit_fill_RSB
authorAndi Kleen <ak@linux.intel.com>
Wed, 17 Jan 2018 22:53:28 +0000 (14:53 -0800)
committerBen Hutchings <ben@decadent.org.uk>
Mon, 19 Mar 2018 18:58:34 +0000 (18:58 +0000)
commit 3f7d875566d8e79c5e0b2c9a413e91b2c29e0854 upstream.

The generated assembler for the C fill RSB inline asm operations has
several issues:

- The C code sets up the loop register, which is then immediately
  overwritten in __FILL_RETURN_BUFFER with the same value again.

- The C code also passes in the iteration count in another register, which
  is not used at all.

Remove these two unnecessary operations. Just rely on the single constant
passed to the macro for the iterations.

Signed-off-by: Andi Kleen <ak@linux.intel.com>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: David Woodhouse <dwmw@amazon.co.uk>
Cc: dave.hansen@intel.com
Cc: gregkh@linuxfoundation.org
Cc: torvalds@linux-foundation.org
Cc: arjan@linux.intel.com
Link: https://lkml.kernel.org/r/20180117225328.15414-1-andi@firstfloor.org
[bwh: Backported to 3.2: adjust contex]
Signed-off-by: Ben Hutchings <ben@decadent.org.uk>
arch/x86/include/asm/nospec-branch.h

index 0fbd082..6b6bdcd 100644 (file)
@@ -184,15 +184,16 @@ extern char __indirect_thunk_size[];
 static inline void vmexit_fill_RSB(void)
 {
 #ifdef CONFIG_RETPOLINE
 static inline void vmexit_fill_RSB(void)
 {
 #ifdef CONFIG_RETPOLINE
-       unsigned long loops = RSB_CLEAR_LOOPS / 2;
+       unsigned long loops;
 
        asm volatile (ALTERNATIVE("jmp 910f",
                                  __stringify(__FILL_RETURN_BUFFER(%0, RSB_CLEAR_LOOPS, %1)),
                                  X86_FEATURE_RETPOLINE)
                      "910:"
 
        asm volatile (ALTERNATIVE("jmp 910f",
                                  __stringify(__FILL_RETURN_BUFFER(%0, RSB_CLEAR_LOOPS, %1)),
                                  X86_FEATURE_RETPOLINE)
                      "910:"
-                     : "=&r" (loops), ASM_CALL_CONSTRAINT
-                     : "r" (loops) : "memory" );
+                     : "=r" (loops), ASM_CALL_CONSTRAINT
+                     : : "memory" );
 #endif
 }
 #endif
 }
+
 #endif /* __ASSEMBLY__ */
 #endif /* __NOSPEC_BRANCH_H__ */
 #endif /* __ASSEMBLY__ */
 #endif /* __NOSPEC_BRANCH_H__ */