mirror of
https://github.com/qemu/qemu.git
synced 2024-11-26 12:23:36 +08:00
target-arm: raise exception on misaligned LDREX operands
Qemu does not generally perform alignment checks. However, the ARM ARM requires implementation of alignment exceptions for a number of cases including LDREX, and Windows-on-ARM relies on this. This change adds plumbing to enable alignment checks on loads using MO_ALIGN, a do_unaligned_access hook to raise the exception (data abort), and uses the new aligned loads in LDREX (for all but single-byte loads). Signed-off-by: Andrew Baumann <Andrew.Baumann@microsoft.com> Message-id: 1449167808-5656-1-git-send-email-Andrew.Baumann@microsoft.com [PMM: set WnR bits in syndrome and FSR as appropriate] Reviewed-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
parent
580106df5f
commit
30901475b9
@ -1417,6 +1417,7 @@ static void arm_cpu_class_init(ObjectClass *oc, void *data)
|
||||
cc->handle_mmu_fault = arm_cpu_handle_mmu_fault;
|
||||
#else
|
||||
cc->do_interrupt = arm_cpu_do_interrupt;
|
||||
cc->do_unaligned_access = arm_cpu_do_unaligned_access;
|
||||
cc->get_phys_page_debug = arm_cpu_get_phys_page_debug;
|
||||
cc->vmsd = &vmstate_arm_cpu;
|
||||
cc->virtio_is_big_endian = arm_cpu_is_big_endian;
|
||||
|
@ -5996,6 +5996,14 @@ static inline bool regime_using_lpae_format(CPUARMState *env,
|
||||
return false;
|
||||
}
|
||||
|
||||
/* Returns true if the translation regime is using LPAE format page tables.
|
||||
* Used when raising alignment exceptions, whose FSR changes depending on
|
||||
* whether the long or short descriptor format is in use. */
|
||||
bool arm_regime_using_lpae_format(CPUARMState *env, ARMMMUIdx mmu_idx)
|
||||
{
|
||||
return regime_using_lpae_format(env, mmu_idx);
|
||||
}
|
||||
|
||||
static inline bool regime_is_user(CPUARMState *env, ARMMMUIdx mmu_idx)
|
||||
{
|
||||
switch (mmu_idx) {
|
||||
|
@ -441,4 +441,11 @@ struct ARMMMUFaultInfo {
|
||||
bool arm_tlb_fill(CPUState *cpu, vaddr address, int rw, int mmu_idx,
|
||||
uint32_t *fsr, ARMMMUFaultInfo *fi);
|
||||
|
||||
/* Return true if the translation regime is using LPAE format page tables */
|
||||
bool arm_regime_using_lpae_format(CPUARMState *env, ARMMMUIdx mmu_idx);
|
||||
|
||||
/* Raise a data fault alignment exception for the specified virtual address */
|
||||
void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr, int is_write,
|
||||
int is_user, uintptr_t retaddr);
|
||||
|
||||
#endif
|
||||
|
@ -126,7 +126,45 @@ void tlb_fill(CPUState *cs, target_ulong addr, int is_write, int mmu_idx,
|
||||
raise_exception(env, exc, syn, target_el);
|
||||
}
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Raise a data fault alignment exception for the specified virtual address */
|
||||
void arm_cpu_do_unaligned_access(CPUState *cs, vaddr vaddr, int is_write,
|
||||
int is_user, uintptr_t retaddr)
|
||||
{
|
||||
ARMCPU *cpu = ARM_CPU(cs);
|
||||
CPUARMState *env = &cpu->env;
|
||||
int target_el;
|
||||
bool same_el;
|
||||
|
||||
if (retaddr) {
|
||||
/* now we have a real cpu fault */
|
||||
cpu_restore_state(cs, retaddr);
|
||||
}
|
||||
|
||||
target_el = exception_target_el(env);
|
||||
same_el = (arm_current_el(env) == target_el);
|
||||
|
||||
env->exception.vaddress = vaddr;
|
||||
|
||||
/* the DFSR for an alignment fault depends on whether we're using
|
||||
* the LPAE long descriptor format, or the short descriptor format
|
||||
*/
|
||||
if (arm_regime_using_lpae_format(env, cpu_mmu_index(env, false))) {
|
||||
env->exception.fsr = 0x21;
|
||||
} else {
|
||||
env->exception.fsr = 0x1;
|
||||
}
|
||||
|
||||
if (is_write == 1 && arm_feature(env, ARM_FEATURE_V6)) {
|
||||
env->exception.fsr |= (1 << 11);
|
||||
}
|
||||
|
||||
raise_exception(env, EXCP_DATA_ABORT,
|
||||
syn_data_abort(same_el, 0, 0, 0, is_write == 1, 0x21),
|
||||
target_el);
|
||||
}
|
||||
|
||||
#endif /* !defined(CONFIG_USER_ONLY) */
|
||||
|
||||
uint32_t HELPER(add_setq)(CPUARMState *env, uint32_t a, uint32_t b)
|
||||
{
|
||||
|
@ -926,13 +926,13 @@ static inline void store_reg_from_load(DisasContext *s, int reg, TCGv_i32 var)
|
||||
#define DO_GEN_LD(SUFF, OPC) \
|
||||
static inline void gen_aa32_ld##SUFF(TCGv_i32 val, TCGv_i32 addr, int index) \
|
||||
{ \
|
||||
tcg_gen_qemu_ld_i32(val, addr, index, OPC); \
|
||||
tcg_gen_qemu_ld_i32(val, addr, index, (OPC)); \
|
||||
}
|
||||
|
||||
#define DO_GEN_ST(SUFF, OPC) \
|
||||
static inline void gen_aa32_st##SUFF(TCGv_i32 val, TCGv_i32 addr, int index) \
|
||||
{ \
|
||||
tcg_gen_qemu_st_i32(val, addr, index, OPC); \
|
||||
tcg_gen_qemu_st_i32(val, addr, index, (OPC)); \
|
||||
}
|
||||
|
||||
static inline void gen_aa32_ld64(TCGv_i64 val, TCGv_i32 addr, int index)
|
||||
@ -988,6 +988,9 @@ DO_GEN_LD(8u, MO_UB)
|
||||
DO_GEN_LD(16s, MO_TESW)
|
||||
DO_GEN_LD(16u, MO_TEUW)
|
||||
DO_GEN_LD(32u, MO_TEUL)
|
||||
/* 'a' variants include an alignment check */
|
||||
DO_GEN_LD(16ua, MO_TEUW | MO_ALIGN)
|
||||
DO_GEN_LD(32ua, MO_TEUL | MO_ALIGN)
|
||||
DO_GEN_ST(8, MO_UB)
|
||||
DO_GEN_ST(16, MO_TEUW)
|
||||
DO_GEN_ST(32, MO_TEUL)
|
||||
@ -7435,11 +7438,11 @@ static void gen_load_exclusive(DisasContext *s, int rt, int rt2,
|
||||
gen_aa32_ld8u(tmp, addr, get_mem_index(s));
|
||||
break;
|
||||
case 1:
|
||||
gen_aa32_ld16u(tmp, addr, get_mem_index(s));
|
||||
gen_aa32_ld16ua(tmp, addr, get_mem_index(s));
|
||||
break;
|
||||
case 2:
|
||||
case 3:
|
||||
gen_aa32_ld32u(tmp, addr, get_mem_index(s));
|
||||
gen_aa32_ld32ua(tmp, addr, get_mem_index(s));
|
||||
break;
|
||||
default:
|
||||
abort();
|
||||
|
Loading…
Reference in New Issue
Block a user