File tree Expand file tree Collapse file tree 3 files changed +10
-9
lines changed Expand file tree Collapse file tree 3 files changed +10
-9
lines changed Original file line number Diff line number Diff line change 17
17
18
18
.pushsection .noinstr.text, "ax"
19
19
20
- SYM_FUNC_START(entry_ibpb)
20
+ /* Clobbers AX, CX, DX */
21
+ SYM_FUNC_START(write_ibpb)
21
22
ANNOTATE_NOENDBR
22
23
movl $MSR_IA32_PRED_CMD, %ecx
23
24
movl $PRED_CMD_IBPB, %eax
@@ -27,9 +28,9 @@ SYM_FUNC_START(entry_ibpb)
27
28
/* Make sure IBPB clears return stack preductions too. */
28
29
FILL_RETURN_BUFFER %rax , RSB_CLEAR_LOOPS, X86_BUG_IBPB_NO_RET
29
30
RET
30
- SYM_FUNC_END(entry_ibpb )
31
+ SYM_FUNC_END(write_ibpb )
31
32
/* For KVM */
32
- EXPORT_SYMBOL_GPL(entry_ibpb );
33
+ EXPORT_SYMBOL_GPL(write_ibpb );
33
34
34
35
.popsection
35
36
Original file line number Diff line number Diff line change 269
269
* typically has NO_MELTDOWN).
270
270
*
271
271
* While retbleed_untrain_ret() doesn't clobber anything but requires stack,
272
- * entry_ibpb () will clobber AX, CX, DX.
272
+ * write_ibpb () will clobber AX, CX, DX.
273
273
*
274
274
* As such, this must be placed after every *SWITCH_TO_KERNEL_CR3 at a point
275
275
* where we have a stack but before any RET instruction.
279
279
VALIDATE_UNRET_END
280
280
CALL_UNTRAIN_RET
281
281
ALTERNATIVE_2 "" , \
282
- "call entry_ibpb " , \ibpb_feature , \
282
+ "call write_ibpb " , \ibpb_feature , \
283
283
__stringify (\call_depth_insns ), X86_FEATURE_CALL_DEPTH
284
284
#endif
285
285
.endm
@@ -368,7 +368,7 @@ extern void srso_return_thunk(void);
368
368
extern void srso_alias_return_thunk (void );
369
369
370
370
extern void entry_untrain_ret (void );
371
- extern void entry_ibpb (void );
371
+ extern void write_ibpb (void );
372
372
373
373
#ifdef CONFIG_X86_64
374
374
extern void clear_bhb_loop (void );
Original file line number Diff line number Diff line change @@ -1142,7 +1142,7 @@ static void __init retbleed_select_mitigation(void)
1142
1142
setup_clear_cpu_cap (X86_FEATURE_RETHUNK );
1143
1143
1144
1144
/*
1145
- * There is no need for RSB filling: entry_ibpb () ensures
1145
+ * There is no need for RSB filling: write_ibpb () ensures
1146
1146
* all predictions, including the RSB, are invalidated,
1147
1147
* regardless of IBPB implementation.
1148
1148
*/
@@ -2676,7 +2676,7 @@ static void __init srso_select_mitigation(void)
2676
2676
setup_clear_cpu_cap (X86_FEATURE_RETHUNK );
2677
2677
2678
2678
/*
2679
- * There is no need for RSB filling: entry_ibpb () ensures
2679
+ * There is no need for RSB filling: write_ibpb () ensures
2680
2680
* all predictions, including the RSB, are invalidated,
2681
2681
* regardless of IBPB implementation.
2682
2682
*/
@@ -2701,7 +2701,7 @@ static void __init srso_select_mitigation(void)
2701
2701
srso_mitigation = SRSO_MITIGATION_IBPB_ON_VMEXIT ;
2702
2702
2703
2703
/*
2704
- * There is no need for RSB filling: entry_ibpb () ensures
2704
+ * There is no need for RSB filling: write_ibpb () ensures
2705
2705
* all predictions, including the RSB, are invalidated,
2706
2706
* regardless of IBPB implementation.
2707
2707
*/
You can’t perform that action at this time.
0 commit comments