mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-18 17:54:13 +08:00
706276543b
Switch to using relative exception table entries on x86. On i386, this has the advantage that the exception table entries don't need to be relocated; on x86-64 this means the exception table entries take up only half the space. In either case, a 32-bit delta is sufficient, as the range of kernel code addresses is limited. Since part of the goal is to avoid needing to adjust the entries when the kernel is relocated, the old trick of using addresses in the NULL pointer range to indicate uaccess_err no longer works (and unlike RISC architectures we can't use a flag bit); instead use an delta just below +2G to indicate these special entries. The reach is still limited to a single instruction. Signed-off-by: H. Peter Anvin <hpa@zytor.com> Cc: David Daney <david.daney@cavium.com> Link: http://lkml.kernel.org/r/CA%2B55aFyijf43qSu3N9nWHEBwaGbb7T2Oq9A=9EyR=Jtyqfq_cQ@mail.gmail.com
170 lines
3.9 KiB
C
170 lines
3.9 KiB
C
#include <linux/module.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/sort.h>
|
|
#include <asm/uaccess.h>
|
|
|
|
static inline unsigned long
|
|
ex_insn_addr(const struct exception_table_entry *x)
|
|
{
|
|
return (unsigned long)&x->insn + x->insn;
|
|
}
|
|
static inline unsigned long
|
|
ex_fixup_addr(const struct exception_table_entry *x)
|
|
{
|
|
return (unsigned long)&x->fixup + x->fixup;
|
|
}
|
|
|
|
int fixup_exception(struct pt_regs *regs)
|
|
{
|
|
const struct exception_table_entry *fixup;
|
|
unsigned long new_ip;
|
|
|
|
#ifdef CONFIG_PNPBIOS
|
|
if (unlikely(SEGMENT_IS_PNP_CODE(regs->cs))) {
|
|
extern u32 pnp_bios_fault_eip, pnp_bios_fault_esp;
|
|
extern u32 pnp_bios_is_utter_crap;
|
|
pnp_bios_is_utter_crap = 1;
|
|
printk(KERN_CRIT "PNPBIOS fault.. attempting recovery.\n");
|
|
__asm__ volatile(
|
|
"movl %0, %%esp\n\t"
|
|
"jmp *%1\n\t"
|
|
: : "g" (pnp_bios_fault_esp), "g" (pnp_bios_fault_eip));
|
|
panic("do_trap: can't hit this");
|
|
}
|
|
#endif
|
|
|
|
fixup = search_exception_tables(regs->ip);
|
|
if (fixup) {
|
|
new_ip = ex_fixup_addr(fixup);
|
|
|
|
if (fixup->fixup - fixup->insn >= 0x7ffffff0 - 4) {
|
|
/* Special hack for uaccess_err */
|
|
current_thread_info()->uaccess_err = 1;
|
|
new_ip -= 0x7ffffff0;
|
|
}
|
|
regs->ip = new_ip;
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Restricted version used during very early boot */
|
|
int __init early_fixup_exception(unsigned long *ip)
|
|
{
|
|
const struct exception_table_entry *fixup;
|
|
unsigned long new_ip;
|
|
|
|
fixup = search_exception_tables(*ip);
|
|
if (fixup) {
|
|
new_ip = ex_fixup_addr(fixup);
|
|
|
|
if (fixup->fixup - fixup->insn >= 0x7ffffff0 - 4) {
|
|
/* uaccess handling not supported during early boot */
|
|
return 0;
|
|
}
|
|
|
|
*ip = new_ip;
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Search one exception table for an entry corresponding to the
|
|
* given instruction address, and return the address of the entry,
|
|
* or NULL if none is found.
|
|
* We use a binary search, and thus we assume that the table is
|
|
* already sorted.
|
|
*/
|
|
const struct exception_table_entry *
|
|
search_extable(const struct exception_table_entry *first,
|
|
const struct exception_table_entry *last,
|
|
unsigned long value)
|
|
{
|
|
while (first <= last) {
|
|
const struct exception_table_entry *mid;
|
|
unsigned long addr;
|
|
|
|
mid = ((last - first) >> 1) + first;
|
|
addr = ex_insn_addr(mid);
|
|
if (addr < value)
|
|
first = mid + 1;
|
|
else if (addr > value)
|
|
last = mid - 1;
|
|
else
|
|
return mid;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* The exception table needs to be sorted so that the binary
|
|
* search that we use to find entries in it works properly.
|
|
* This is used both for the kernel exception table and for
|
|
* the exception tables of modules that get loaded.
|
|
*
|
|
*/
|
|
static int cmp_ex(const void *a, const void *b)
|
|
{
|
|
const struct exception_table_entry *x = a, *y = b;
|
|
|
|
/*
|
|
* This value will always end up fittin in an int, because on
|
|
* both i386 and x86-64 the kernel symbol-reachable address
|
|
* space is < 2 GiB.
|
|
*
|
|
* This compare is only valid after normalization.
|
|
*/
|
|
return x->insn - y->insn;
|
|
}
|
|
|
|
void sort_extable(struct exception_table_entry *start,
|
|
struct exception_table_entry *finish)
|
|
{
|
|
struct exception_table_entry *p;
|
|
int i;
|
|
|
|
/* Convert all entries to being relative to the start of the section */
|
|
i = 0;
|
|
for (p = start; p < finish; p++) {
|
|
p->insn += i;
|
|
i += 4;
|
|
p->fixup += i;
|
|
i += 4;
|
|
}
|
|
|
|
sort(start, finish - start, sizeof(struct exception_table_entry),
|
|
cmp_ex, NULL);
|
|
|
|
/* Denormalize all entries */
|
|
i = 0;
|
|
for (p = start; p < finish; p++) {
|
|
p->insn -= i;
|
|
i += 4;
|
|
p->fixup -= i;
|
|
i += 4;
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_MODULES
|
|
/*
|
|
* If the exception table is sorted, any referring to the module init
|
|
* will be at the beginning or the end.
|
|
*/
|
|
void trim_init_extable(struct module *m)
|
|
{
|
|
/*trim the beginning*/
|
|
while (m->num_exentries &&
|
|
within_module_init(ex_insn_addr(&m->extable[0]), m)) {
|
|
m->extable++;
|
|
m->num_exentries--;
|
|
}
|
|
/*trim the end*/
|
|
while (m->num_exentries &&
|
|
within_module_init(ex_insn_addr(&m->extable[m->num_exentries-1]), m))
|
|
m->num_exentries--;
|
|
}
|
|
#endif /* CONFIG_MODULES */
|