We've so far relied on a patching infrastructure that only gave us a single alternative, without any way to finely control what gets patched. For a single feature, this is an all or nothing thing. It would be interesting to have a more fine grained way of patching the kernel though, where we could dynamically tune the code that gets injected. In order to achive this, let's introduce a new form of alternative that is associated with a callback. This callback gets the instruction sequence number and the old instruction as a parameter, and returns the new instruction. This callback is always called, as the patching decision is now done at runtime (not patching is equivalent to returning the same instruction). Patching with a callback is declared with the new ALTERNATIVE_CB and alternative_cb directives: asm volatile(ALTERNATIVE_CB("mov %0, #0\n", callback) : "r" (v)); or alternative_cb callback mov x0, #0 alternative_else_nop_endif where callback is the C function computing the alternative. Signed-off-by: Marc Zyngier <marc.zyngier@xxxxxxx> --- arch/arm64/include/asm/alternative.h | 34 +++++++++++++++++++++--------- arch/arm64/include/asm/alternative_types.h | 4 ++++ arch/arm64/kernel/alternative.c | 13 ++++++++++-- 3 files changed, 39 insertions(+), 12 deletions(-) diff --git a/arch/arm64/include/asm/alternative.h b/arch/arm64/include/asm/alternative.h index 395befde7595..8a8740a03514 100644 --- a/arch/arm64/include/asm/alternative.h +++ b/arch/arm64/include/asm/alternative.h @@ -18,8 +18,9 @@ void __init apply_alternatives_all(void); void apply_alternatives(void *start, size_t length); -#define ALTINSTR_ENTRY(feature) \ +#define ALTINSTR_ENTRY(feature,cb) \ " .align " __stringify(ALTINSTR_ALIGN) "\n" \ + " .quad " __stringify(cb) "\n" /* callback */ \ " .word 661b - .\n" /* label */ \ " .word 663f - .\n" /* new instruction */ \ " .hword " __stringify(feature) "\n" /* feature bit */ \ @@ -40,13 +41,13 @@ void apply_alternatives(void *start, size_t length); * be fixed in a binutils release posterior to 2.25.51.0.2 (anything * containing commit 4e4d08cf7399b606 or c1baaddf8861). */ -#define __ALTERNATIVE_CFG(oldinstr, newinstr, feature, cfg_enabled) \ +#define __ALTERNATIVE_CFG(oldinstr, newinstr, feature, cfg_enabled, cb) \ ".if "__stringify(cfg_enabled)" == 1\n" \ "661:\n\t" \ oldinstr "\n" \ "662:\n" \ ".pushsection .altinstructions,\"a\"\n" \ - ALTINSTR_ENTRY(feature) \ + ALTINSTR_ENTRY(feature,cb) \ ".popsection\n" \ ".pushsection .altinstr_replacement, \"a\"\n" \ "663:\n\t" \ @@ -58,14 +59,17 @@ void apply_alternatives(void *start, size_t length); ".endif\n" #define _ALTERNATIVE_CFG(oldinstr, newinstr, feature, cfg, ...) \ - __ALTERNATIVE_CFG(oldinstr, newinstr, feature, IS_ENABLED(cfg)) + __ALTERNATIVE_CFG(oldinstr, newinstr, feature, IS_ENABLED(cfg), 0) +#define _ALTERNATIVE_CB(oldinstr, cb, ...) \ + __ALTERNATIVE_CFG(oldinstr, oldinstr, ARM64_NCAPS, 1, cb) #else #include <asm/assembler.h> -.macro altinstruction_entry orig_offset alt_offset feature orig_len alt_len +.macro altinstruction_entry orig_offset, alt_offset, feature, orig_len, alt_len, cb = 0 .align ALTINSTR_ALIGN + .quad \cb .word \orig_offset - . .word \alt_offset - . .hword \feature @@ -73,11 +77,11 @@ void apply_alternatives(void *start, size_t length); .byte \alt_len .endm -.macro alternative_insn insn1, insn2, cap, enable = 1 +.macro alternative_insn insn1, insn2, cap, enable = 1, cb = 0 .if \enable 661: \insn1 662: .pushsection .altinstructions, "a" - altinstruction_entry 661b, 663f, \cap, 662b-661b, 664f-663f + altinstruction_entry 661b, 663f, \cap, 662b-661b, 664f-663f, \cb .popsection .pushsection .altinstr_replacement, "ax" 663: \insn2 @@ -109,10 +113,10 @@ void apply_alternatives(void *start, size_t length); /* * Begin an alternative code sequence. */ -.macro alternative_if_not cap +.macro alternative_if_not cap, cb = 0 .set .Lasm_alt_mode, 0 .pushsection .altinstructions, "a" - altinstruction_entry 661f, 663f, \cap, 662f-661f, 664f-663f + altinstruction_entry 661f, 663f, \cap, 662f-661f, 664f-663f, \cb .popsection 661: .endm @@ -120,13 +124,17 @@ void apply_alternatives(void *start, size_t length); .macro alternative_if cap .set .Lasm_alt_mode, 1 .pushsection .altinstructions, "a" - altinstruction_entry 663f, 661f, \cap, 664f-663f, 662f-661f + altinstruction_entry 663f, 661f, \cap, 664f-663f, 662f-661f, 0 .popsection .pushsection .altinstr_replacement, "ax" .align 2 /* So GAS knows label 661 is suitably aligned */ 661: .endm +.macro alternative_cb cb + alternative_if_not ARM64_NCAPS, \cb +.endm + /* * Provide the other half of the alternative code sequence. */ @@ -166,6 +174,9 @@ alternative_endif #define _ALTERNATIVE_CFG(insn1, insn2, cap, cfg, ...) \ alternative_insn insn1, insn2, cap, IS_ENABLED(cfg) +#define _ALTERNATIVE_CB(insn1, cb, ...) \ + alternative_insn insn1, insn1, ARM64_NCAPS, 1, cb + .macro user_alt, label, oldinstr, newinstr, cond 9999: alternative_insn "\oldinstr", "\newinstr", \cond _ASM_EXTABLE 9999b, \label @@ -242,4 +253,7 @@ alternative_endif #define ALTERNATIVE(oldinstr, newinstr, ...) \ _ALTERNATIVE_CFG(oldinstr, newinstr, __VA_ARGS__, 1) +#define ALTERNATIVE_CB(oldinstr, cb, ...) \ + _ALTERNATIVE_CB(oldinstr, cb) + #endif /* __ASM_ALTERNATIVE_H */ diff --git a/arch/arm64/include/asm/alternative_types.h b/arch/arm64/include/asm/alternative_types.h index 26cf76167f2d..7ac90d69602e 100644 --- a/arch/arm64/include/asm/alternative_types.h +++ b/arch/arm64/include/asm/alternative_types.h @@ -2,7 +2,11 @@ #ifndef __ASM_ALTERNATIVE_TYPES_H #define __ASM_ALTERNATIVE_TYPES_H +struct alt_instr; +typedef u32 (*alternative_cb_t)(struct alt_instr *alt, int index, u32 new_insn); + struct alt_instr { + alternative_cb_t cb; /* Callback for dynamic patching */ s32 orig_offset; /* offset to original instruction */ s32 alt_offset; /* offset to replacement instruction */ u16 cpufeature; /* cpufeature bit set for replacement */ diff --git a/arch/arm64/kernel/alternative.c b/arch/arm64/kernel/alternative.c index 6dd0a3a3e5c9..7812d537be25 100644 --- a/arch/arm64/kernel/alternative.c +++ b/arch/arm64/kernel/alternative.c @@ -115,9 +115,12 @@ static void __apply_alternatives(void *alt_region, bool use_linear_alias) u32 insn; int i, nr_inst; - if (!cpus_have_cap(alt->cpufeature)) + /* Use ARM64_NCAPS as an unconditional patch */ + if (alt->cpufeature != ARM64_NCAPS && + !cpus_have_cap(alt->cpufeature)) continue; + BUG_ON(alt->cpufeature == ARM64_NCAPS && !alt->cb); BUG_ON(alt->alt_len != alt->orig_len); pr_info_once("patching kernel code\n"); @@ -128,7 +131,13 @@ static void __apply_alternatives(void *alt_region, bool use_linear_alias) nr_inst = alt->alt_len / sizeof(insn); for (i = 0; i < nr_inst; i++) { - insn = get_alt_insn(alt, origptr + i, replptr + i); + if (alt->cb) { + insn = le32_to_cpu(updptr[i]); + insn = alt->cb(alt, i, insn); + } else { + insn = get_alt_insn(alt, origptr + i, + replptr + i); + } updptr[i] = cpu_to_le32(insn); } -- 2.14.2