1 // SPDX-License-Identifier: GPL-2.0
3 * jump label x86 support
5 * Copyright (C) 2009 Jason Baron <jbaron@redhat.com>
8 #include <linux/jump_label.h>
9 #include <linux/memory.h>
10 #include <linux/uaccess.h>
11 #include <linux/module.h>
12 #include <linux/list.h>
13 #include <linux/jhash.h>
14 #include <linux/cpu.h>
15 #include <asm/kprobes.h>
16 #include <asm/alternative.h>
17 #include <asm/text-patching.h>
20 int arch_jump_entry_size(struct jump_entry
*entry
)
22 struct insn insn
= {};
24 insn_decode_kernel(&insn
, (void *)jump_entry_code(entry
));
25 BUG_ON(insn
.length
!= 2 && insn
.length
!= 5);
30 struct jump_label_patch
{
35 static struct jump_label_patch
36 __jump_label_patch(struct jump_entry
*entry
, enum jump_label_type type
)
38 const void *expect
, *code
, *nop
;
39 const void *addr
, *dest
;
42 addr
= (void *)jump_entry_code(entry
);
43 dest
= (void *)jump_entry_target(entry
);
45 size
= arch_jump_entry_size(entry
);
48 code
= text_gen_insn(JMP8_INSN_OPCODE
, addr
, dest
);
53 code
= text_gen_insn(JMP32_INSN_OPCODE
, addr
, dest
);
60 if (type
== JUMP_LABEL_JMP
)
65 if (memcmp(addr
, expect
, size
)) {
67 * The location is not an op that we were expecting.
68 * Something went wrong. Crash the box, as something could be
69 * corrupting the kernel.
71 pr_crit("jump_label: Fatal kernel bug, unexpected op at %pS [%p] (%5ph != %5ph)) size:%d type:%d\n",
72 addr
, addr
, addr
, expect
, size
, type
);
76 if (type
== JUMP_LABEL_NOP
)
79 return (struct jump_label_patch
){.code
= code
, .size
= size
};
82 static __always_inline
void
83 __jump_label_transform(struct jump_entry
*entry
,
84 enum jump_label_type type
,
87 const struct jump_label_patch jlp
= __jump_label_patch(entry
, type
);
90 * As long as only a single processor is running and the code is still
91 * not marked as RO, text_poke_early() can be used; Checking that
92 * system_state is SYSTEM_BOOTING guarantees it. It will be set to
93 * SYSTEM_SCHEDULING before other cores are awaken and before the
94 * code is write-protected.
96 * At the time the change is being done, just ignore whether we
97 * are doing nop -> jump or jump -> nop transition, and assume
98 * always nop being the 'currently valid' instruction
100 if (init
|| system_state
== SYSTEM_BOOTING
) {
101 text_poke_early((void *)jump_entry_code(entry
), jlp
.code
, jlp
.size
);
105 text_poke_bp((void *)jump_entry_code(entry
), jlp
.code
, jlp
.size
, NULL
);
108 static void __ref
jump_label_transform(struct jump_entry
*entry
,
109 enum jump_label_type type
,
112 mutex_lock(&text_mutex
);
113 __jump_label_transform(entry
, type
, init
);
114 mutex_unlock(&text_mutex
);
117 void arch_jump_label_transform(struct jump_entry
*entry
,
118 enum jump_label_type type
)
120 jump_label_transform(entry
, type
, 0);
123 bool arch_jump_label_transform_queue(struct jump_entry
*entry
,
124 enum jump_label_type type
)
126 struct jump_label_patch jlp
;
128 if (system_state
== SYSTEM_BOOTING
) {
130 * Fallback to the non-batching mode.
132 arch_jump_label_transform(entry
, type
);
136 mutex_lock(&text_mutex
);
137 jlp
= __jump_label_patch(entry
, type
);
138 text_poke_queue((void *)jump_entry_code(entry
), jlp
.code
, jlp
.size
, NULL
);
139 mutex_unlock(&text_mutex
);
143 void arch_jump_label_transform_apply(void)
145 mutex_lock(&text_mutex
);
147 mutex_unlock(&text_mutex
);