mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-28 15:13:55 +08:00
ba54f0c3f7
Currently, the jump label of a static key is transformed via the arch specific function: void arch_jump_label_transform(struct jump_entry *entry, enum jump_label_type type) The new approach (batch mode) uses two arch functions, the first has the same arguments of the arch_jump_label_transform(), and is the function: bool arch_jump_label_transform_queue(struct jump_entry *entry, enum jump_label_type type) Rather than transforming the code, it adds the jump_entry in a queue of entries to be updated. This functions returns true in the case of a successful enqueue of an entry. If it returns false, the caller must to apply the queue and then try to queue again, for instance, because the queue is full. This function expects the caller to sort the entries by the address before enqueueuing then. This is already done by the arch independent code, though. After queuing all jump_entries, the function: void arch_jump_label_transform_apply(void) Applies the changes in the queue. Signed-off-by: Daniel Bristot de Oliveira <bristot@redhat.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Borislav Petkov <bp@alien8.de> Cc: Chris von Recklinghausen <crecklin@redhat.com> Cc: Clark Williams <williams@redhat.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: H. Peter Anvin <hpa@zytor.com> Cc: Jason Baron <jbaron@akamai.com> Cc: Jiri Kosina <jkosina@suse.cz> Cc: Josh Poimboeuf <jpoimboe@redhat.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Marcelo Tosatti <mtosatti@redhat.com> Cc: Masami Hiramatsu <mhiramat@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Scott Wood <swood@redhat.com> Cc: Steven Rostedt (VMware) <rostedt@goodmis.org> Cc: Thomas Gleixner <tglx@linutronix.de> Link: https://lkml.kernel.org/r/57b4caa654bad7e3b066301c9a9ae233dea065b5.1560325897.git.bristot@redhat.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
201 lines
5.3 KiB
C
201 lines
5.3 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* jump label x86 support
|
|
*
|
|
* Copyright (C) 2009 Jason Baron <jbaron@redhat.com>
|
|
*
|
|
*/
|
|
#include <linux/jump_label.h>
|
|
#include <linux/memory.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/module.h>
|
|
#include <linux/list.h>
|
|
#include <linux/jhash.h>
|
|
#include <linux/cpu.h>
|
|
#include <asm/kprobes.h>
|
|
#include <asm/alternative.h>
|
|
#include <asm/text-patching.h>
|
|
|
|
union jump_code_union {
|
|
char code[JUMP_LABEL_NOP_SIZE];
|
|
struct {
|
|
char jump;
|
|
int offset;
|
|
} __attribute__((packed));
|
|
};
|
|
|
|
static void bug_at(unsigned char *ip, int line)
|
|
{
|
|
/*
|
|
* The location is not an op that we were expecting.
|
|
* Something went wrong. Crash the box, as something could be
|
|
* corrupting the kernel.
|
|
*/
|
|
pr_crit("jump_label: Fatal kernel bug, unexpected op at %pS [%p] (%5ph) %d\n", ip, ip, ip, line);
|
|
BUG();
|
|
}
|
|
|
|
static void __jump_label_set_jump_code(struct jump_entry *entry,
|
|
enum jump_label_type type,
|
|
union jump_code_union *code,
|
|
int init)
|
|
{
|
|
const unsigned char default_nop[] = { STATIC_KEY_INIT_NOP };
|
|
const unsigned char *ideal_nop = ideal_nops[NOP_ATOMIC5];
|
|
const void *expect;
|
|
int line;
|
|
|
|
code->jump = 0xe9;
|
|
code->offset = jump_entry_target(entry) -
|
|
(jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE);
|
|
|
|
if (init) {
|
|
expect = default_nop; line = __LINE__;
|
|
} else if (type == JUMP_LABEL_JMP) {
|
|
expect = ideal_nop; line = __LINE__;
|
|
} else {
|
|
expect = code->code; line = __LINE__;
|
|
}
|
|
|
|
if (memcmp((void *)jump_entry_code(entry), expect, JUMP_LABEL_NOP_SIZE))
|
|
bug_at((void *)jump_entry_code(entry), line);
|
|
|
|
if (type == JUMP_LABEL_NOP)
|
|
memcpy(code, ideal_nop, JUMP_LABEL_NOP_SIZE);
|
|
}
|
|
|
|
static void __ref __jump_label_transform(struct jump_entry *entry,
|
|
enum jump_label_type type,
|
|
int init)
|
|
{
|
|
union jump_code_union code;
|
|
|
|
__jump_label_set_jump_code(entry, type, &code, init);
|
|
|
|
/*
|
|
* As long as only a single processor is running and the code is still
|
|
* not marked as RO, text_poke_early() can be used; Checking that
|
|
* system_state is SYSTEM_BOOTING guarantees it. It will be set to
|
|
* SYSTEM_SCHEDULING before other cores are awaken and before the
|
|
* code is write-protected.
|
|
*
|
|
* At the time the change is being done, just ignore whether we
|
|
* are doing nop -> jump or jump -> nop transition, and assume
|
|
* always nop being the 'currently valid' instruction
|
|
*/
|
|
if (init || system_state == SYSTEM_BOOTING) {
|
|
text_poke_early((void *)jump_entry_code(entry), &code,
|
|
JUMP_LABEL_NOP_SIZE);
|
|
return;
|
|
}
|
|
|
|
text_poke_bp((void *)jump_entry_code(entry), &code, JUMP_LABEL_NOP_SIZE,
|
|
(void *)jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE);
|
|
}
|
|
|
|
void arch_jump_label_transform(struct jump_entry *entry,
|
|
enum jump_label_type type)
|
|
{
|
|
mutex_lock(&text_mutex);
|
|
__jump_label_transform(entry, type, 0);
|
|
mutex_unlock(&text_mutex);
|
|
}
|
|
|
|
#define TP_VEC_MAX (PAGE_SIZE / sizeof(struct text_poke_loc))
|
|
static struct text_poke_loc tp_vec[TP_VEC_MAX];
|
|
int tp_vec_nr = 0;
|
|
|
|
bool arch_jump_label_transform_queue(struct jump_entry *entry,
|
|
enum jump_label_type type)
|
|
{
|
|
struct text_poke_loc *tp;
|
|
void *entry_code;
|
|
|
|
if (system_state == SYSTEM_BOOTING) {
|
|
/*
|
|
* Fallback to the non-batching mode.
|
|
*/
|
|
arch_jump_label_transform(entry, type);
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* No more space in the vector, tell upper layer to apply
|
|
* the queue before continuing.
|
|
*/
|
|
if (tp_vec_nr == TP_VEC_MAX)
|
|
return false;
|
|
|
|
tp = &tp_vec[tp_vec_nr];
|
|
|
|
entry_code = (void *)jump_entry_code(entry);
|
|
|
|
/*
|
|
* The INT3 handler will do a bsearch in the queue, so we need entries
|
|
* to be sorted. We can survive an unsorted list by rejecting the entry,
|
|
* forcing the generic jump_label code to apply the queue. Warning once,
|
|
* to raise the attention to the case of an unsorted entry that is
|
|
* better not happen, because, in the worst case we will perform in the
|
|
* same way as we do without batching - with some more overhead.
|
|
*/
|
|
if (tp_vec_nr > 0) {
|
|
int prev = tp_vec_nr - 1;
|
|
struct text_poke_loc *prev_tp = &tp_vec[prev];
|
|
|
|
if (WARN_ON_ONCE(prev_tp->addr > entry_code))
|
|
return false;
|
|
}
|
|
|
|
__jump_label_set_jump_code(entry, type,
|
|
(union jump_code_union *) &tp->opcode, 0);
|
|
|
|
tp->addr = entry_code;
|
|
tp->detour = entry_code + JUMP_LABEL_NOP_SIZE;
|
|
tp->len = JUMP_LABEL_NOP_SIZE;
|
|
|
|
tp_vec_nr++;
|
|
|
|
return true;
|
|
}
|
|
|
|
void arch_jump_label_transform_apply(void)
|
|
{
|
|
if (!tp_vec_nr)
|
|
return;
|
|
|
|
mutex_lock(&text_mutex);
|
|
text_poke_bp_batch(tp_vec, tp_vec_nr);
|
|
mutex_unlock(&text_mutex);
|
|
|
|
tp_vec_nr = 0;
|
|
}
|
|
|
|
static enum {
|
|
JL_STATE_START,
|
|
JL_STATE_NO_UPDATE,
|
|
JL_STATE_UPDATE,
|
|
} jlstate __initdata_or_module = JL_STATE_START;
|
|
|
|
__init_or_module void arch_jump_label_transform_static(struct jump_entry *entry,
|
|
enum jump_label_type type)
|
|
{
|
|
/*
|
|
* This function is called at boot up and when modules are
|
|
* first loaded. Check if the default nop, the one that is
|
|
* inserted at compile time, is the ideal nop. If it is, then
|
|
* we do not need to update the nop, and we can leave it as is.
|
|
* If it is not, then we need to update the nop to the ideal nop.
|
|
*/
|
|
if (jlstate == JL_STATE_START) {
|
|
const unsigned char default_nop[] = { STATIC_KEY_INIT_NOP };
|
|
const unsigned char *ideal_nop = ideal_nops[NOP_ATOMIC5];
|
|
|
|
if (memcmp(ideal_nop, default_nop, 5) != 0)
|
|
jlstate = JL_STATE_UPDATE;
|
|
else
|
|
jlstate = JL_STATE_NO_UPDATE;
|
|
}
|
|
if (jlstate == JL_STATE_UPDATE)
|
|
__jump_label_transform(entry, type, 1);
|
|
}
|