binutils-gdb/gdb/gdbserver/linux-ppc-low.c
Joel Brobecker e2882c8578 Update copyright year range in all GDB files
gdb/ChangeLog:

        Update copyright year range in all GDB files
2018-01-02 07:38:06 +04:00

3170 lines
79 KiB
C

/* GNU/Linux/PowerPC specific low level interface, for the remote server for
GDB.
Copyright (C) 1995-2018 Free Software Foundation, Inc.
This file is part of GDB.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>. */
#include "server.h"
#include "linux-low.h"
#include <elf.h>
#include <asm/ptrace.h>
#include "nat/ppc-linux.h"
#include "linux-ppc-tdesc.h"
#include "ax.h"
#include "tracepoint.h"
#define PPC_FIELD(value, from, len) \
(((value) >> (32 - (from) - (len))) & ((1 << (len)) - 1))
#define PPC_SEXT(v, bs) \
((((CORE_ADDR) (v) & (((CORE_ADDR) 1 << (bs)) - 1)) \
^ ((CORE_ADDR) 1 << ((bs) - 1))) \
- ((CORE_ADDR) 1 << ((bs) - 1)))
#define PPC_OP6(insn) PPC_FIELD (insn, 0, 6)
#define PPC_BO(insn) PPC_FIELD (insn, 6, 5)
#define PPC_LI(insn) (PPC_SEXT (PPC_FIELD (insn, 6, 24), 24) << 2)
#define PPC_BD(insn) (PPC_SEXT (PPC_FIELD (insn, 16, 14), 14) << 2)
static unsigned long ppc_hwcap;
#define ppc_num_regs 73
#ifdef __powerpc64__
/* We use a constant for FPSCR instead of PT_FPSCR, because
many shipped PPC64 kernels had the wrong value in ptrace.h. */
static int ppc_regmap[] =
{PT_R0 * 8, PT_R1 * 8, PT_R2 * 8, PT_R3 * 8,
PT_R4 * 8, PT_R5 * 8, PT_R6 * 8, PT_R7 * 8,
PT_R8 * 8, PT_R9 * 8, PT_R10 * 8, PT_R11 * 8,
PT_R12 * 8, PT_R13 * 8, PT_R14 * 8, PT_R15 * 8,
PT_R16 * 8, PT_R17 * 8, PT_R18 * 8, PT_R19 * 8,
PT_R20 * 8, PT_R21 * 8, PT_R22 * 8, PT_R23 * 8,
PT_R24 * 8, PT_R25 * 8, PT_R26 * 8, PT_R27 * 8,
PT_R28 * 8, PT_R29 * 8, PT_R30 * 8, PT_R31 * 8,
PT_FPR0*8, PT_FPR0*8 + 8, PT_FPR0*8+16, PT_FPR0*8+24,
PT_FPR0*8+32, PT_FPR0*8+40, PT_FPR0*8+48, PT_FPR0*8+56,
PT_FPR0*8+64, PT_FPR0*8+72, PT_FPR0*8+80, PT_FPR0*8+88,
PT_FPR0*8+96, PT_FPR0*8+104, PT_FPR0*8+112, PT_FPR0*8+120,
PT_FPR0*8+128, PT_FPR0*8+136, PT_FPR0*8+144, PT_FPR0*8+152,
PT_FPR0*8+160, PT_FPR0*8+168, PT_FPR0*8+176, PT_FPR0*8+184,
PT_FPR0*8+192, PT_FPR0*8+200, PT_FPR0*8+208, PT_FPR0*8+216,
PT_FPR0*8+224, PT_FPR0*8+232, PT_FPR0*8+240, PT_FPR0*8+248,
PT_NIP * 8, PT_MSR * 8, PT_CCR * 8, PT_LNK * 8,
PT_CTR * 8, PT_XER * 8, PT_FPR0*8 + 256,
PT_ORIG_R3 * 8, PT_TRAP * 8 };
#else
/* Currently, don't check/send MQ. */
static int ppc_regmap[] =
{PT_R0 * 4, PT_R1 * 4, PT_R2 * 4, PT_R3 * 4,
PT_R4 * 4, PT_R5 * 4, PT_R6 * 4, PT_R7 * 4,
PT_R8 * 4, PT_R9 * 4, PT_R10 * 4, PT_R11 * 4,
PT_R12 * 4, PT_R13 * 4, PT_R14 * 4, PT_R15 * 4,
PT_R16 * 4, PT_R17 * 4, PT_R18 * 4, PT_R19 * 4,
PT_R20 * 4, PT_R21 * 4, PT_R22 * 4, PT_R23 * 4,
PT_R24 * 4, PT_R25 * 4, PT_R26 * 4, PT_R27 * 4,
PT_R28 * 4, PT_R29 * 4, PT_R30 * 4, PT_R31 * 4,
PT_FPR0*4, PT_FPR0*4 + 8, PT_FPR0*4+16, PT_FPR0*4+24,
PT_FPR0*4+32, PT_FPR0*4+40, PT_FPR0*4+48, PT_FPR0*4+56,
PT_FPR0*4+64, PT_FPR0*4+72, PT_FPR0*4+80, PT_FPR0*4+88,
PT_FPR0*4+96, PT_FPR0*4+104, PT_FPR0*4+112, PT_FPR0*4+120,
PT_FPR0*4+128, PT_FPR0*4+136, PT_FPR0*4+144, PT_FPR0*4+152,
PT_FPR0*4+160, PT_FPR0*4+168, PT_FPR0*4+176, PT_FPR0*4+184,
PT_FPR0*4+192, PT_FPR0*4+200, PT_FPR0*4+208, PT_FPR0*4+216,
PT_FPR0*4+224, PT_FPR0*4+232, PT_FPR0*4+240, PT_FPR0*4+248,
PT_NIP * 4, PT_MSR * 4, PT_CCR * 4, PT_LNK * 4,
PT_CTR * 4, PT_XER * 4, PT_FPSCR * 4,
PT_ORIG_R3 * 4, PT_TRAP * 4
};
static int ppc_regmap_e500[] =
{PT_R0 * 4, PT_R1 * 4, PT_R2 * 4, PT_R3 * 4,
PT_R4 * 4, PT_R5 * 4, PT_R6 * 4, PT_R7 * 4,
PT_R8 * 4, PT_R9 * 4, PT_R10 * 4, PT_R11 * 4,
PT_R12 * 4, PT_R13 * 4, PT_R14 * 4, PT_R15 * 4,
PT_R16 * 4, PT_R17 * 4, PT_R18 * 4, PT_R19 * 4,
PT_R20 * 4, PT_R21 * 4, PT_R22 * 4, PT_R23 * 4,
PT_R24 * 4, PT_R25 * 4, PT_R26 * 4, PT_R27 * 4,
PT_R28 * 4, PT_R29 * 4, PT_R30 * 4, PT_R31 * 4,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
-1, -1, -1, -1,
PT_NIP * 4, PT_MSR * 4, PT_CCR * 4, PT_LNK * 4,
PT_CTR * 4, PT_XER * 4, -1,
PT_ORIG_R3 * 4, PT_TRAP * 4
};
#endif
static int
ppc_cannot_store_register (int regno)
{
const struct target_desc *tdesc = current_process ()->tdesc;
#ifndef __powerpc64__
/* Some kernels do not allow us to store fpscr. */
if (!(ppc_hwcap & PPC_FEATURE_HAS_SPE)
&& regno == find_regno (tdesc, "fpscr"))
return 2;
#endif
/* Some kernels do not allow us to store orig_r3 or trap. */
if (regno == find_regno (tdesc, "orig_r3")
|| regno == find_regno (tdesc, "trap"))
return 2;
return 0;
}
static int
ppc_cannot_fetch_register (int regno)
{
return 0;
}
static void
ppc_collect_ptrace_register (struct regcache *regcache, int regno, char *buf)
{
memset (buf, 0, sizeof (long));
if (__BYTE_ORDER == __LITTLE_ENDIAN)
{
/* Little-endian values always sit at the left end of the buffer. */
collect_register (regcache, regno, buf);
}
else if (__BYTE_ORDER == __BIG_ENDIAN)
{
/* Big-endian values sit at the right end of the buffer. In case of
registers whose sizes are smaller than sizeof (long), we must use a
padding to access them correctly. */
int size = register_size (regcache->tdesc, regno);
if (size < sizeof (long))
collect_register (regcache, regno, buf + sizeof (long) - size);
else
collect_register (regcache, regno, buf);
}
else
perror_with_name ("Unexpected byte order");
}
static void
ppc_supply_ptrace_register (struct regcache *regcache,
int regno, const char *buf)
{
if (__BYTE_ORDER == __LITTLE_ENDIAN)
{
/* Little-endian values always sit at the left end of the buffer. */
supply_register (regcache, regno, buf);
}
else if (__BYTE_ORDER == __BIG_ENDIAN)
{
/* Big-endian values sit at the right end of the buffer. In case of
registers whose sizes are smaller than sizeof (long), we must use a
padding to access them correctly. */
int size = register_size (regcache->tdesc, regno);
if (size < sizeof (long))
supply_register (regcache, regno, buf + sizeof (long) - size);
else
supply_register (regcache, regno, buf);
}
else
perror_with_name ("Unexpected byte order");
}
#define INSTR_SC 0x44000002
#define NR_spu_run 0x0116
/* If the PPU thread is currently stopped on a spu_run system call,
return to FD and ADDR the file handle and NPC parameter address
used with the system call. Return non-zero if successful. */
static int
parse_spufs_run (struct regcache *regcache, int *fd, CORE_ADDR *addr)
{
CORE_ADDR curr_pc;
int curr_insn;
int curr_r0;
if (register_size (regcache->tdesc, 0) == 4)
{
unsigned int pc, r0, r3, r4;
collect_register_by_name (regcache, "pc", &pc);
collect_register_by_name (regcache, "r0", &r0);
collect_register_by_name (regcache, "orig_r3", &r3);
collect_register_by_name (regcache, "r4", &r4);
curr_pc = (CORE_ADDR) pc;
curr_r0 = (int) r0;
*fd = (int) r3;
*addr = (CORE_ADDR) r4;
}
else
{
unsigned long pc, r0, r3, r4;
collect_register_by_name (regcache, "pc", &pc);
collect_register_by_name (regcache, "r0", &r0);
collect_register_by_name (regcache, "orig_r3", &r3);
collect_register_by_name (regcache, "r4", &r4);
curr_pc = (CORE_ADDR) pc;
curr_r0 = (int) r0;
*fd = (int) r3;
*addr = (CORE_ADDR) r4;
}
/* Fetch instruction preceding current NIP. */
if ((*the_target->read_memory) (curr_pc - 4,
(unsigned char *) &curr_insn, 4) != 0)
return 0;
/* It should be a "sc" instruction. */
if (curr_insn != INSTR_SC)
return 0;
/* System call number should be NR_spu_run. */
if (curr_r0 != NR_spu_run)
return 0;
return 1;
}
static CORE_ADDR
ppc_get_pc (struct regcache *regcache)
{
CORE_ADDR addr;
int fd;
if (parse_spufs_run (regcache, &fd, &addr))
{
unsigned int pc;
(*the_target->read_memory) (addr, (unsigned char *) &pc, 4);
return ((CORE_ADDR)1 << 63)
| ((CORE_ADDR)fd << 32) | (CORE_ADDR) (pc - 4);
}
else if (register_size (regcache->tdesc, 0) == 4)
{
unsigned int pc;
collect_register_by_name (regcache, "pc", &pc);
return (CORE_ADDR) pc;
}
else
{
unsigned long pc;
collect_register_by_name (regcache, "pc", &pc);
return (CORE_ADDR) pc;
}
}
static void
ppc_set_pc (struct regcache *regcache, CORE_ADDR pc)
{
CORE_ADDR addr;
int fd;
if (parse_spufs_run (regcache, &fd, &addr))
{
unsigned int newpc = pc;
(*the_target->write_memory) (addr, (unsigned char *) &newpc, 4);
}
else if (register_size (regcache->tdesc, 0) == 4)
{
unsigned int newpc = pc;
supply_register_by_name (regcache, "pc", &newpc);
}
else
{
unsigned long newpc = pc;
supply_register_by_name (regcache, "pc", &newpc);
}
}
static int
ppc_get_auxv (unsigned long type, unsigned long *valp)
{
const struct target_desc *tdesc = current_process ()->tdesc;
int wordsize = register_size (tdesc, 0);
unsigned char *data = (unsigned char *) alloca (2 * wordsize);
int offset = 0;
while ((*the_target->read_auxv) (offset, data, 2 * wordsize) == 2 * wordsize)
{
if (wordsize == 4)
{
unsigned int *data_p = (unsigned int *)data;
if (data_p[0] == type)
{
*valp = data_p[1];
return 1;
}
}
else
{
unsigned long *data_p = (unsigned long *)data;
if (data_p[0] == type)
{
*valp = data_p[1];
return 1;
}
}
offset += 2 * wordsize;
}
*valp = 0;
return 0;
}
#ifndef __powerpc64__
static int ppc_regmap_adjusted;
#endif
/* Correct in either endianness.
This instruction is "twge r2, r2", which GDB uses as a software
breakpoint. */
static const unsigned int ppc_breakpoint = 0x7d821008;
#define ppc_breakpoint_len 4
/* Implementation of linux_target_ops method "sw_breakpoint_from_kind". */
static const gdb_byte *
ppc_sw_breakpoint_from_kind (int kind, int *size)
{
*size = ppc_breakpoint_len;
return (const gdb_byte *) &ppc_breakpoint;
}
static int
ppc_breakpoint_at (CORE_ADDR where)
{
unsigned int insn;
if (where & ((CORE_ADDR)1 << 63))
{
char mem_annex[32];
sprintf (mem_annex, "%d/mem", (int)((where >> 32) & 0x7fffffff));
(*the_target->qxfer_spu) (mem_annex, (unsigned char *) &insn,
NULL, where & 0xffffffff, 4);
if (insn == 0x3fff)
return 1;
}
else
{
(*the_target->read_memory) (where, (unsigned char *) &insn, 4);
if (insn == ppc_breakpoint)
return 1;
/* If necessary, recognize more trap instructions here. GDB only uses
the one. */
}
return 0;
}
/* Implement supports_z_point_type target-ops.
Returns true if type Z_TYPE breakpoint is supported.
Handling software breakpoint at server side, so tracepoints
and breakpoints can be inserted at the same location. */
static int
ppc_supports_z_point_type (char z_type)
{
switch (z_type)
{
case Z_PACKET_SW_BP:
return 1;
case Z_PACKET_HW_BP:
case Z_PACKET_WRITE_WP:
case Z_PACKET_ACCESS_WP:
default:
return 0;
}
}
/* Implement insert_point target-ops.
Returns 0 on success, -1 on failure and 1 on unsupported. */
static int
ppc_insert_point (enum raw_bkpt_type type, CORE_ADDR addr,
int size, struct raw_breakpoint *bp)
{
switch (type)
{
case raw_bkpt_type_sw:
return insert_memory_breakpoint (bp);
case raw_bkpt_type_hw:
case raw_bkpt_type_write_wp:
case raw_bkpt_type_access_wp:
default:
/* Unsupported. */
return 1;
}
}
/* Implement remove_point target-ops.
Returns 0 on success, -1 on failure and 1 on unsupported. */
static int
ppc_remove_point (enum raw_bkpt_type type, CORE_ADDR addr,
int size, struct raw_breakpoint *bp)
{
switch (type)
{
case raw_bkpt_type_sw:
return remove_memory_breakpoint (bp);
case raw_bkpt_type_hw:
case raw_bkpt_type_write_wp:
case raw_bkpt_type_access_wp:
default:
/* Unsupported. */
return 1;
}
}
/* Provide only a fill function for the general register set. ps_lgetregs
will use this for NPTL support. */
static void ppc_fill_gregset (struct regcache *regcache, void *buf)
{
int i;
for (i = 0; i < 32; i++)
ppc_collect_ptrace_register (regcache, i, (char *) buf + ppc_regmap[i]);
for (i = 64; i < 70; i++)
ppc_collect_ptrace_register (regcache, i, (char *) buf + ppc_regmap[i]);
for (i = 71; i < 73; i++)
ppc_collect_ptrace_register (regcache, i, (char *) buf + ppc_regmap[i]);
}
#define SIZEOF_VSXREGS 32*8
static void
ppc_fill_vsxregset (struct regcache *regcache, void *buf)
{
int i, base;
char *regset = (char *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_VSX))
return;
base = find_regno (regcache->tdesc, "vs0h");
for (i = 0; i < 32; i++)
collect_register (regcache, base + i, &regset[i * 8]);
}
static void
ppc_store_vsxregset (struct regcache *regcache, const void *buf)
{
int i, base;
const char *regset = (const char *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_VSX))
return;
base = find_regno (regcache->tdesc, "vs0h");
for (i = 0; i < 32; i++)
supply_register (regcache, base + i, &regset[i * 8]);
}
#define SIZEOF_VRREGS 33*16+4
static void
ppc_fill_vrregset (struct regcache *regcache, void *buf)
{
int i, base;
char *regset = (char *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_ALTIVEC))
return;
base = find_regno (regcache->tdesc, "vr0");
for (i = 0; i < 32; i++)
collect_register (regcache, base + i, &regset[i * 16]);
collect_register_by_name (regcache, "vscr", &regset[32 * 16 + 12]);
collect_register_by_name (regcache, "vrsave", &regset[33 * 16]);
}
static void
ppc_store_vrregset (struct regcache *regcache, const void *buf)
{
int i, base;
const char *regset = (const char *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_ALTIVEC))
return;
base = find_regno (regcache->tdesc, "vr0");
for (i = 0; i < 32; i++)
supply_register (regcache, base + i, &regset[i * 16]);
supply_register_by_name (regcache, "vscr", &regset[32 * 16 + 12]);
supply_register_by_name (regcache, "vrsave", &regset[33 * 16]);
}
struct gdb_evrregset_t
{
unsigned long evr[32];
unsigned long long acc;
unsigned long spefscr;
};
static void
ppc_fill_evrregset (struct regcache *regcache, void *buf)
{
int i, ev0;
struct gdb_evrregset_t *regset = (struct gdb_evrregset_t *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_SPE))
return;
ev0 = find_regno (regcache->tdesc, "ev0h");
for (i = 0; i < 32; i++)
collect_register (regcache, ev0 + i, &regset->evr[i]);
collect_register_by_name (regcache, "acc", &regset->acc);
collect_register_by_name (regcache, "spefscr", &regset->spefscr);
}
static void
ppc_store_evrregset (struct regcache *regcache, const void *buf)
{
int i, ev0;
const struct gdb_evrregset_t *regset = (const struct gdb_evrregset_t *) buf;
if (!(ppc_hwcap & PPC_FEATURE_HAS_SPE))
return;
ev0 = find_regno (regcache->tdesc, "ev0h");
for (i = 0; i < 32; i++)
supply_register (regcache, ev0 + i, &regset->evr[i]);
supply_register_by_name (regcache, "acc", &regset->acc);
supply_register_by_name (regcache, "spefscr", &regset->spefscr);
}
/* Support for hardware single step. */
static int
ppc_supports_hardware_single_step (void)
{
return 1;
}
static struct regset_info ppc_regsets[] = {
/* List the extra register sets before GENERAL_REGS. That way we will
fetch them every time, but still fall back to PTRACE_PEEKUSER for the
general registers. Some kernels support these, but not the newer
PPC_PTRACE_GETREGS. */
{ PTRACE_GETVSXREGS, PTRACE_SETVSXREGS, 0, SIZEOF_VSXREGS, EXTENDED_REGS,
ppc_fill_vsxregset, ppc_store_vsxregset },
{ PTRACE_GETVRREGS, PTRACE_SETVRREGS, 0, SIZEOF_VRREGS, EXTENDED_REGS,
ppc_fill_vrregset, ppc_store_vrregset },
{ PTRACE_GETEVRREGS, PTRACE_SETEVRREGS, 0, 32 * 4 + 8 + 4, EXTENDED_REGS,
ppc_fill_evrregset, ppc_store_evrregset },
{ 0, 0, 0, 0, GENERAL_REGS, ppc_fill_gregset, NULL },
NULL_REGSET
};
static struct usrregs_info ppc_usrregs_info =
{
ppc_num_regs,
ppc_regmap,
};
static struct regsets_info ppc_regsets_info =
{
ppc_regsets, /* regsets */
0, /* num_regsets */
NULL, /* disabled_regsets */
};
static struct regs_info regs_info =
{
NULL, /* regset_bitmap */
&ppc_usrregs_info,
&ppc_regsets_info
};
static const struct regs_info *
ppc_regs_info (void)
{
return &regs_info;
}
static void
ppc_arch_setup (void)
{
const struct target_desc *tdesc;
#ifdef __powerpc64__
long msr;
struct regcache *regcache;
/* On a 64-bit host, assume 64-bit inferior process with no
AltiVec registers. Reset ppc_hwcap to ensure that the
collect_register call below does not fail. */
tdesc = tdesc_powerpc_64l;
current_process ()->tdesc = tdesc;
ppc_hwcap = 0;
regcache = new_register_cache (tdesc);
fetch_inferior_registers (regcache, find_regno (tdesc, "msr"));
collect_register_by_name (regcache, "msr", &msr);
free_register_cache (regcache);
if (ppc64_64bit_inferior_p (msr))
{
ppc_get_auxv (AT_HWCAP, &ppc_hwcap);
if (ppc_hwcap & PPC_FEATURE_CELL)
tdesc = tdesc_powerpc_cell64l;
else if (ppc_hwcap & PPC_FEATURE_HAS_VSX)
{
/* Power ISA 2.05 (implemented by Power 6 and newer processors)
increases the FPSCR from 32 bits to 64 bits. Even though Power 7
supports this ISA version, it doesn't have PPC_FEATURE_ARCH_2_05
set, only PPC_FEATURE_ARCH_2_06. Since for now the only bits
used in the higher half of the register are for Decimal Floating
Point, we check if that feature is available to decide the size
of the FPSCR. */
if (ppc_hwcap & PPC_FEATURE_HAS_DFP)
tdesc = tdesc_powerpc_isa205_vsx64l;
else
tdesc = tdesc_powerpc_vsx64l;
}
else if (ppc_hwcap & PPC_FEATURE_HAS_ALTIVEC)
{
if (ppc_hwcap & PPC_FEATURE_HAS_DFP)
tdesc = tdesc_powerpc_isa205_altivec64l;
else
tdesc = tdesc_powerpc_altivec64l;
}
current_process ()->tdesc = tdesc;
return;
}
#endif
/* OK, we have a 32-bit inferior. */
tdesc = tdesc_powerpc_32l;
current_process ()->tdesc = tdesc;
ppc_get_auxv (AT_HWCAP, &ppc_hwcap);
if (ppc_hwcap & PPC_FEATURE_CELL)
tdesc = tdesc_powerpc_cell32l;
else if (ppc_hwcap & PPC_FEATURE_HAS_VSX)
{
if (ppc_hwcap & PPC_FEATURE_HAS_DFP)
tdesc = tdesc_powerpc_isa205_vsx32l;
else
tdesc = tdesc_powerpc_vsx32l;
}
else if (ppc_hwcap & PPC_FEATURE_HAS_ALTIVEC)
{
if (ppc_hwcap & PPC_FEATURE_HAS_DFP)
tdesc = tdesc_powerpc_isa205_altivec32l;
else
tdesc = tdesc_powerpc_altivec32l;
}
/* On 32-bit machines, check for SPE registers.
Set the low target's regmap field as appropriately. */
#ifndef __powerpc64__
if (ppc_hwcap & PPC_FEATURE_HAS_SPE)
tdesc = tdesc_powerpc_e500l;
if (!ppc_regmap_adjusted)
{
if (ppc_hwcap & PPC_FEATURE_HAS_SPE)
ppc_usrregs_info.regmap = ppc_regmap_e500;
/* If the FPSCR is 64-bit wide, we need to fetch the whole
64-bit slot and not just its second word. The PT_FPSCR
supplied in a 32-bit GDB compilation doesn't reflect
this. */
if (register_size (tdesc, 70) == 8)
ppc_regmap[70] = (48 + 2*32) * sizeof (long);
ppc_regmap_adjusted = 1;
}
#endif
current_process ()->tdesc = tdesc;
}
/* Implementation of linux_target_ops method "supports_tracepoints". */
static int
ppc_supports_tracepoints (void)
{
return 1;
}
/* Get the thread area address. This is used to recognize which
thread is which when tracing with the in-process agent library. We
don't read anything from the address, and treat it as opaque; it's
the address itself that we assume is unique per-thread. */
static int
ppc_get_thread_area (int lwpid, CORE_ADDR *addr)
{
struct lwp_info *lwp = find_lwp_pid (pid_to_ptid (lwpid));
struct thread_info *thr = get_lwp_thread (lwp);
struct regcache *regcache = get_thread_regcache (thr, 1);
ULONGEST tp = 0;
#ifdef __powerpc64__
if (register_size (regcache->tdesc, 0) == 8)
collect_register_by_name (regcache, "r13", &tp);
else
#endif
collect_register_by_name (regcache, "r2", &tp);
*addr = tp;
return 0;
}
#ifdef __powerpc64__
/* Older glibc doesn't provide this. */
#ifndef EF_PPC64_ABI
#define EF_PPC64_ABI 3
#endif
/* Returns 1 if inferior is using ELFv2 ABI. Undefined for 32-bit
inferiors. */
static int
is_elfv2_inferior (void)
{
/* To be used as fallback if we're unable to determine the right result -
assume inferior uses the same ABI as gdbserver. */
#if _CALL_ELF == 2
const int def_res = 1;
#else
const int def_res = 0;
#endif
unsigned long phdr;
Elf64_Ehdr ehdr;
if (!ppc_get_auxv (AT_PHDR, &phdr))
return def_res;
/* Assume ELF header is at the beginning of the page where program headers
are located. If it doesn't look like one, bail. */
read_inferior_memory (phdr & ~0xfff, (unsigned char *) &ehdr, sizeof ehdr);
if (memcmp(ehdr.e_ident, ELFMAG, SELFMAG))
return def_res;
return (ehdr.e_flags & EF_PPC64_ABI) == 2;
}
#endif
/* Generate a ds-form instruction in BUF and return the number of bytes written
0 6 11 16 30 32
| OPCD | RST | RA | DS |XO| */
__attribute__((unused)) /* Maybe unused due to conditional compilation. */
static int
gen_ds_form (uint32_t *buf, int opcd, int rst, int ra, int ds, int xo)
{
uint32_t insn;
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((rst & ~0x1f) == 0);
gdb_assert ((ra & ~0x1f) == 0);
gdb_assert ((xo & ~0x3) == 0);
insn = (rst << 21) | (ra << 16) | (ds & 0xfffc) | (xo & 0x3);
*buf = (opcd << 26) | insn;
return 1;
}
/* Followings are frequently used ds-form instructions. */
#define GEN_STD(buf, rs, ra, offset) gen_ds_form (buf, 62, rs, ra, offset, 0)
#define GEN_STDU(buf, rs, ra, offset) gen_ds_form (buf, 62, rs, ra, offset, 1)
#define GEN_LD(buf, rt, ra, offset) gen_ds_form (buf, 58, rt, ra, offset, 0)
#define GEN_LDU(buf, rt, ra, offset) gen_ds_form (buf, 58, rt, ra, offset, 1)
/* Generate a d-form instruction in BUF.
0 6 11 16 32
| OPCD | RST | RA | D | */
static int
gen_d_form (uint32_t *buf, int opcd, int rst, int ra, int si)
{
uint32_t insn;
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((rst & ~0x1f) == 0);
gdb_assert ((ra & ~0x1f) == 0);
insn = (rst << 21) | (ra << 16) | (si & 0xffff);
*buf = (opcd << 26) | insn;
return 1;
}
/* Followings are frequently used d-form instructions. */
#define GEN_ADDI(buf, rt, ra, si) gen_d_form (buf, 14, rt, ra, si)
#define GEN_ADDIS(buf, rt, ra, si) gen_d_form (buf, 15, rt, ra, si)
#define GEN_LI(buf, rt, si) GEN_ADDI (buf, rt, 0, si)
#define GEN_LIS(buf, rt, si) GEN_ADDIS (buf, rt, 0, si)
#define GEN_ORI(buf, rt, ra, si) gen_d_form (buf, 24, rt, ra, si)
#define GEN_ORIS(buf, rt, ra, si) gen_d_form (buf, 25, rt, ra, si)
#define GEN_LWZ(buf, rt, ra, si) gen_d_form (buf, 32, rt, ra, si)
#define GEN_STW(buf, rt, ra, si) gen_d_form (buf, 36, rt, ra, si)
#define GEN_STWU(buf, rt, ra, si) gen_d_form (buf, 37, rt, ra, si)
/* Generate a xfx-form instruction in BUF and return the number of bytes
written.
0 6 11 21 31 32
| OPCD | RST | RI | XO |/| */
static int
gen_xfx_form (uint32_t *buf, int opcd, int rst, int ri, int xo)
{
uint32_t insn;
unsigned int n = ((ri & 0x1f) << 5) | ((ri >> 5) & 0x1f);
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((rst & ~0x1f) == 0);
gdb_assert ((xo & ~0x3ff) == 0);
insn = (rst << 21) | (n << 11) | (xo << 1);
*buf = (opcd << 26) | insn;
return 1;
}
/* Followings are frequently used xfx-form instructions. */
#define GEN_MFSPR(buf, rt, spr) gen_xfx_form (buf, 31, rt, spr, 339)
#define GEN_MTSPR(buf, rt, spr) gen_xfx_form (buf, 31, rt, spr, 467)
#define GEN_MFCR(buf, rt) gen_xfx_form (buf, 31, rt, 0, 19)
#define GEN_MTCR(buf, rt) gen_xfx_form (buf, 31, rt, 0x3cf, 144)
#define GEN_SYNC(buf, L, E) gen_xfx_form (buf, 31, L & 0x3, \
E & 0xf, 598)
#define GEN_LWSYNC(buf) GEN_SYNC (buf, 1, 0)
/* Generate a x-form instruction in BUF and return the number of bytes written.
0 6 11 16 21 31 32
| OPCD | RST | RA | RB | XO |RC| */
static int
gen_x_form (uint32_t *buf, int opcd, int rst, int ra, int rb, int xo, int rc)
{
uint32_t insn;
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((rst & ~0x1f) == 0);
gdb_assert ((ra & ~0x1f) == 0);
gdb_assert ((rb & ~0x1f) == 0);
gdb_assert ((xo & ~0x3ff) == 0);
gdb_assert ((rc & ~1) == 0);
insn = (rst << 21) | (ra << 16) | (rb << 11) | (xo << 1) | rc;
*buf = (opcd << 26) | insn;
return 1;
}
/* Followings are frequently used x-form instructions. */
#define GEN_OR(buf, ra, rs, rb) gen_x_form (buf, 31, rs, ra, rb, 444, 0)
#define GEN_MR(buf, ra, rs) GEN_OR (buf, ra, rs, rs)
#define GEN_LWARX(buf, rt, ra, rb) gen_x_form (buf, 31, rt, ra, rb, 20, 0)
#define GEN_STWCX(buf, rs, ra, rb) gen_x_form (buf, 31, rs, ra, rb, 150, 1)
/* Assume bf = cr7. */
#define GEN_CMPW(buf, ra, rb) gen_x_form (buf, 31, 28, ra, rb, 0, 0)
/* Generate a md-form instruction in BUF and return the number of bytes written.
0 6 11 16 21 27 30 31 32
| OPCD | RS | RA | sh | mb | XO |sh|Rc| */
static int
gen_md_form (uint32_t *buf, int opcd, int rs, int ra, int sh, int mb,
int xo, int rc)
{
uint32_t insn;
unsigned int n = ((mb & 0x1f) << 1) | ((mb >> 5) & 0x1);
unsigned int sh0_4 = sh & 0x1f;
unsigned int sh5 = (sh >> 5) & 1;
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((rs & ~0x1f) == 0);
gdb_assert ((ra & ~0x1f) == 0);
gdb_assert ((sh & ~0x3f) == 0);
gdb_assert ((mb & ~0x3f) == 0);
gdb_assert ((xo & ~0x7) == 0);
gdb_assert ((rc & ~0x1) == 0);
insn = (rs << 21) | (ra << 16) | (sh0_4 << 11) | (n << 5)
| (sh5 << 1) | (xo << 2) | (rc & 1);
*buf = (opcd << 26) | insn;
return 1;
}
/* The following are frequently used md-form instructions. */
#define GEN_RLDICL(buf, ra, rs ,sh, mb) \
gen_md_form (buf, 30, rs, ra, sh, mb, 0, 0)
#define GEN_RLDICR(buf, ra, rs ,sh, mb) \
gen_md_form (buf, 30, rs, ra, sh, mb, 1, 0)
/* Generate a i-form instruction in BUF and return the number of bytes written.
0 6 30 31 32
| OPCD | LI |AA|LK| */
static int
gen_i_form (uint32_t *buf, int opcd, int li, int aa, int lk)
{
uint32_t insn;
gdb_assert ((opcd & ~0x3f) == 0);
insn = (li & 0x3fffffc) | (aa & 1) | (lk & 1);
*buf = (opcd << 26) | insn;
return 1;
}
/* The following are frequently used i-form instructions. */
#define GEN_B(buf, li) gen_i_form (buf, 18, li, 0, 0)
#define GEN_BL(buf, li) gen_i_form (buf, 18, li, 0, 1)
/* Generate a b-form instruction in BUF and return the number of bytes written.
0 6 11 16 30 31 32
| OPCD | BO | BI | BD |AA|LK| */
static int
gen_b_form (uint32_t *buf, int opcd, int bo, int bi, int bd,
int aa, int lk)
{
uint32_t insn;
gdb_assert ((opcd & ~0x3f) == 0);
gdb_assert ((bo & ~0x1f) == 0);
gdb_assert ((bi & ~0x1f) == 0);
insn = (bo << 21) | (bi << 16) | (bd & 0xfffc) | (aa & 1) | (lk & 1);
*buf = (opcd << 26) | insn;
return 1;
}
/* The following are frequently used b-form instructions. */
/* Assume bi = cr7. */
#define GEN_BNE(buf, bd) gen_b_form (buf, 16, 0x4, (7 << 2) | 2, bd, 0 ,0)
/* GEN_LOAD and GEN_STORE generate 64- or 32-bit load/store for ppc64 or ppc32
respectively. They are primary used for save/restore GPRs in jump-pad,
not used for bytecode compiling. */
#ifdef __powerpc64__
#define GEN_LOAD(buf, rt, ra, si, is_64) (is_64 ? \
GEN_LD (buf, rt, ra, si) : \
GEN_LWZ (buf, rt, ra, si))
#define GEN_STORE(buf, rt, ra, si, is_64) (is_64 ? \
GEN_STD (buf, rt, ra, si) : \
GEN_STW (buf, rt, ra, si))
#else
#define GEN_LOAD(buf, rt, ra, si, is_64) GEN_LWZ (buf, rt, ra, si)
#define GEN_STORE(buf, rt, ra, si, is_64) GEN_STW (buf, rt, ra, si)
#endif
/* Generate a sequence of instructions to load IMM in the register REG.
Write the instructions in BUF and return the number of bytes written. */
static int
gen_limm (uint32_t *buf, int reg, uint64_t imm, int is_64)
{
uint32_t *p = buf;
if ((imm + 32768) < 65536)
{
/* li reg, imm[15:0] */
p += GEN_LI (p, reg, imm);
}
else if ((imm >> 32) == 0)
{
/* lis reg, imm[31:16]
ori reg, reg, imm[15:0]
rldicl reg, reg, 0, 32 */
p += GEN_LIS (p, reg, (imm >> 16) & 0xffff);
if ((imm & 0xffff) != 0)
p += GEN_ORI (p, reg, reg, imm & 0xffff);
/* Clear upper 32-bit if sign-bit is set. */
if (imm & (1u << 31) && is_64)
p += GEN_RLDICL (p, reg, reg, 0, 32);
}
else
{
gdb_assert (is_64);
/* lis reg, <imm[63:48]>
ori reg, reg, <imm[48:32]>
rldicr reg, reg, 32, 31
oris reg, reg, <imm[31:16]>
ori reg, reg, <imm[15:0]> */
p += GEN_LIS (p, reg, ((imm >> 48) & 0xffff));
if (((imm >> 32) & 0xffff) != 0)
p += GEN_ORI (p, reg, reg, ((imm >> 32) & 0xffff));
p += GEN_RLDICR (p, reg, reg, 32, 31);
if (((imm >> 16) & 0xffff) != 0)
p += GEN_ORIS (p, reg, reg, ((imm >> 16) & 0xffff));
if ((imm & 0xffff) != 0)
p += GEN_ORI (p, reg, reg, (imm & 0xffff));
}
return p - buf;
}
/* Generate a sequence for atomically exchange at location LOCK.
This code sequence clobbers r6, r7, r8. LOCK is the location for
the atomic-xchg, OLD_VALUE is expected old value stored in the
location, and R_NEW is a register for the new value. */
static int
gen_atomic_xchg (uint32_t *buf, CORE_ADDR lock, int old_value, int r_new,
int is_64)
{
const int r_lock = 6;
const int r_old = 7;
const int r_tmp = 8;
uint32_t *p = buf;
/*
1: lwarx TMP, 0, LOCK
cmpwi TMP, OLD
bne 1b
stwcx. NEW, 0, LOCK
bne 1b */
p += gen_limm (p, r_lock, lock, is_64);
p += gen_limm (p, r_old, old_value, is_64);
p += GEN_LWARX (p, r_tmp, 0, r_lock);
p += GEN_CMPW (p, r_tmp, r_old);
p += GEN_BNE (p, -8);
p += GEN_STWCX (p, r_new, 0, r_lock);
p += GEN_BNE (p, -16);
return p - buf;
}
/* Generate a sequence of instructions for calling a function
at address of FN. Return the number of bytes are written in BUF. */
static int
gen_call (uint32_t *buf, CORE_ADDR fn, int is_64, int is_opd)
{
uint32_t *p = buf;
/* Must be called by r12 for caller to calculate TOC address. */
p += gen_limm (p, 12, fn, is_64);
if (is_opd)
{
p += GEN_LOAD (p, 11, 12, 16, is_64);
p += GEN_LOAD (p, 2, 12, 8, is_64);
p += GEN_LOAD (p, 12, 12, 0, is_64);
}
p += GEN_MTSPR (p, 12, 9); /* mtctr r12 */
*p++ = 0x4e800421; /* bctrl */
return p - buf;
}
/* Copy the instruction from OLDLOC to *TO, and update *TO to *TO + size
of instruction. This function is used to adjust pc-relative instructions
when copying. */
static void
ppc_relocate_instruction (CORE_ADDR *to, CORE_ADDR oldloc)
{
uint32_t insn, op6;
long rel, newrel;
read_inferior_memory (oldloc, (unsigned char *) &insn, 4);
op6 = PPC_OP6 (insn);
if (op6 == 18 && (insn & 2) == 0)
{
/* branch && AA = 0 */
rel = PPC_LI (insn);
newrel = (oldloc - *to) + rel;
/* Out of range. Cannot relocate instruction. */
if (newrel >= (1 << 25) || newrel < -(1 << 25))
return;
insn = (insn & ~0x3fffffc) | (newrel & 0x3fffffc);
}
else if (op6 == 16 && (insn & 2) == 0)
{
/* conditional branch && AA = 0 */
/* If the new relocation is too big for even a 26-bit unconditional
branch, there is nothing we can do. Just abort.
Otherwise, if it can be fit in 16-bit conditional branch, just
copy the instruction and relocate the address.
If the it's big for conditional-branch (16-bit), try to invert the
condition and jump with 26-bit branch. For example,
beq .Lgoto
INSN1
=>
bne 1f (+8)
b .Lgoto
1:INSN1
After this transform, we are actually jump from *TO+4 instead of *TO,
so check the relocation again because it will be 1-insn farther then
before if *TO is after OLDLOC.
For BDNZT (or so) is transformed from
bdnzt eq, .Lgoto
INSN1
=>
bdz 1f (+12)
bf eq, 1f (+8)
b .Lgoto
1:INSN1
See also "BO field encodings". */
rel = PPC_BD (insn);
newrel = (oldloc - *to) + rel;
if (newrel < (1 << 15) && newrel >= -(1 << 15))
insn = (insn & ~0xfffc) | (newrel & 0xfffc);
else if ((PPC_BO (insn) & 0x14) == 0x4 || (PPC_BO (insn) & 0x14) == 0x10)
{
newrel -= 4;
/* Out of range. Cannot relocate instruction. */
if (newrel >= (1 << 25) || newrel < -(1 << 25))
return;
if ((PPC_BO (insn) & 0x14) == 0x4)
insn ^= (1 << 24);
else if ((PPC_BO (insn) & 0x14) == 0x10)
insn ^= (1 << 22);
/* Jump over the unconditional branch. */
insn = (insn & ~0xfffc) | 0x8;
write_inferior_memory (*to, (unsigned char *) &insn, 4);
*to += 4;
/* Build a unconditional branch and copy LK bit. */
insn = (18 << 26) | (0x3fffffc & newrel) | (insn & 0x3);
write_inferior_memory (*to, (unsigned char *) &insn, 4);
*to += 4;
return;
}
else if ((PPC_BO (insn) & 0x14) == 0)
{
uint32_t bdnz_insn = (16 << 26) | (0x10 << 21) | 12;
uint32_t bf_insn = (16 << 26) | (0x4 << 21) | 8;
newrel -= 8;
/* Out of range. Cannot relocate instruction. */
if (newrel >= (1 << 25) || newrel < -(1 << 25))
return;
/* Copy BI field. */
bf_insn |= (insn & 0x1f0000);
/* Invert condition. */
bdnz_insn |= (insn ^ (1 << 22)) & (1 << 22);
bf_insn |= (insn ^ (1 << 24)) & (1 << 24);
write_inferior_memory (*to, (unsigned char *) &bdnz_insn, 4);
*to += 4;
write_inferior_memory (*to, (unsigned char *) &bf_insn, 4);
*to += 4;
/* Build a unconditional branch and copy LK bit. */
insn = (18 << 26) | (0x3fffffc & newrel) | (insn & 0x3);
write_inferior_memory (*to, (unsigned char *) &insn, 4);
*to += 4;
return;
}
else /* (BO & 0x14) == 0x14, branch always. */
{
/* Out of range. Cannot relocate instruction. */
if (newrel >= (1 << 25) || newrel < -(1 << 25))
return;
/* Build a unconditional branch and copy LK bit. */
insn = (18 << 26) | (0x3fffffc & newrel) | (insn & 0x3);
write_inferior_memory (*to, (unsigned char *) &insn, 4);
*to += 4;
return;
}
}
write_inferior_memory (*to, (unsigned char *) &insn, 4);
*to += 4;
}
/* Implement install_fast_tracepoint_jump_pad of target_ops.
See target.h for details. */
static int
ppc_install_fast_tracepoint_jump_pad (CORE_ADDR tpoint, CORE_ADDR tpaddr,
CORE_ADDR collector,
CORE_ADDR lockaddr,
ULONGEST orig_size,
CORE_ADDR *jump_entry,
CORE_ADDR *trampoline,
ULONGEST *trampoline_size,
unsigned char *jjump_pad_insn,
ULONGEST *jjump_pad_insn_size,
CORE_ADDR *adjusted_insn_addr,
CORE_ADDR *adjusted_insn_addr_end,
char *err)
{
uint32_t buf[256];
uint32_t *p = buf;
int j, offset;
CORE_ADDR buildaddr = *jump_entry;
const CORE_ADDR entryaddr = *jump_entry;
int rsz, min_frame, frame_size, tp_reg;
#ifdef __powerpc64__
struct regcache *regcache = get_thread_regcache (current_thread, 0);
int is_64 = register_size (regcache->tdesc, 0) == 8;
int is_opd = is_64 && !is_elfv2_inferior ();
#else
int is_64 = 0, is_opd = 0;
#endif
#ifdef __powerpc64__
if (is_64)
{
/* Minimum frame size is 32 bytes for ELFv2, and 112 bytes for ELFv1. */
rsz = 8;
min_frame = 112;
frame_size = (40 * rsz) + min_frame;
tp_reg = 13;
}
else
{
#endif
rsz = 4;
min_frame = 16;
frame_size = (40 * rsz) + min_frame;
tp_reg = 2;
#ifdef __powerpc64__
}
#endif
/* Stack frame layout for this jump pad,
High thread_area (r13/r2) |
tpoint - collecting_t obj
PC/<tpaddr> | +36
CTR | +35
LR | +34
XER | +33
CR | +32
R31 |
R29 |
... |
R1 | +1
R0 - collected registers
... |
... |
Low Back-chain -
The code flow of this jump pad,
1. Adjust SP
2. Save GPR and SPR
3. Prepare argument
4. Call gdb_collector
5. Restore GPR and SPR
6. Restore SP
7. Build a jump for back to the program
8. Copy/relocate original instruction
9. Build a jump for replacing orignal instruction. */
/* Adjust stack pointer. */
if (is_64)
p += GEN_STDU (p, 1, 1, -frame_size); /* stdu r1,-frame_size(r1) */
else
p += GEN_STWU (p, 1, 1, -frame_size); /* stwu r1,-frame_size(r1) */
/* Store GPRs. Save R1 later, because it had just been modified, but
we want the original value. */
for (j = 2; j < 32; j++)
p += GEN_STORE (p, j, 1, min_frame + j * rsz, is_64);
p += GEN_STORE (p, 0, 1, min_frame + 0 * rsz, is_64);
/* Set r0 to the original value of r1 before adjusting stack frame,
and then save it. */
p += GEN_ADDI (p, 0, 1, frame_size);
p += GEN_STORE (p, 0, 1, min_frame + 1 * rsz, is_64);
/* Save CR, XER, LR, and CTR. */
p += GEN_MFCR (p, 3); /* mfcr r3 */
p += GEN_MFSPR (p, 4, 1); /* mfxer r4 */
p += GEN_MFSPR (p, 5, 8); /* mflr r5 */
p += GEN_MFSPR (p, 6, 9); /* mfctr r6 */
p += GEN_STORE (p, 3, 1, min_frame + 32 * rsz, is_64);/* std r3, 32(r1) */
p += GEN_STORE (p, 4, 1, min_frame + 33 * rsz, is_64);/* std r4, 33(r1) */
p += GEN_STORE (p, 5, 1, min_frame + 34 * rsz, is_64);/* std r5, 34(r1) */
p += GEN_STORE (p, 6, 1, min_frame + 35 * rsz, is_64);/* std r6, 35(r1) */
/* Save PC<tpaddr> */
p += gen_limm (p, 3, tpaddr, is_64);
p += GEN_STORE (p, 3, 1, min_frame + 36 * rsz, is_64);
/* Setup arguments to collector. */
/* Set r4 to collected registers. */
p += GEN_ADDI (p, 4, 1, min_frame);
/* Set r3 to TPOINT. */
p += gen_limm (p, 3, tpoint, is_64);
/* Prepare collecting_t object for lock. */
p += GEN_STORE (p, 3, 1, min_frame + 37 * rsz, is_64);
p += GEN_STORE (p, tp_reg, 1, min_frame + 38 * rsz, is_64);
/* Set R5 to collecting object. */
p += GEN_ADDI (p, 5, 1, 37 * rsz);
p += GEN_LWSYNC (p);
p += gen_atomic_xchg (p, lockaddr, 0, 5, is_64);
p += GEN_LWSYNC (p);
/* Call to collector. */
p += gen_call (p, collector, is_64, is_opd);
/* Simply write 0 to release the lock. */
p += gen_limm (p, 3, lockaddr, is_64);
p += gen_limm (p, 4, 0, is_64);
p += GEN_LWSYNC (p);
p += GEN_STORE (p, 4, 3, 0, is_64);
/* Restore stack and registers. */
p += GEN_LOAD (p, 3, 1, min_frame + 32 * rsz, is_64); /* ld r3, 32(r1) */
p += GEN_LOAD (p, 4, 1, min_frame + 33 * rsz, is_64); /* ld r4, 33(r1) */
p += GEN_LOAD (p, 5, 1, min_frame + 34 * rsz, is_64); /* ld r5, 34(r1) */
p += GEN_LOAD (p, 6, 1, min_frame + 35 * rsz, is_64); /* ld r6, 35(r1) */
p += GEN_MTCR (p, 3); /* mtcr r3 */
p += GEN_MTSPR (p, 4, 1); /* mtxer r4 */
p += GEN_MTSPR (p, 5, 8); /* mtlr r5 */
p += GEN_MTSPR (p, 6, 9); /* mtctr r6 */
/* Restore GPRs. */
for (j = 2; j < 32; j++)
p += GEN_LOAD (p, j, 1, min_frame + j * rsz, is_64);
p += GEN_LOAD (p, 0, 1, min_frame + 0 * rsz, is_64);
/* Restore SP. */
p += GEN_ADDI (p, 1, 1, frame_size);
/* Flush instructions to inferior memory. */
write_inferior_memory (buildaddr, (unsigned char *) buf, (p - buf) * 4);
/* Now, insert the original instruction to execute in the jump pad. */
*adjusted_insn_addr = buildaddr + (p - buf) * 4;
*adjusted_insn_addr_end = *adjusted_insn_addr;
ppc_relocate_instruction (adjusted_insn_addr_end, tpaddr);
/* Verify the relocation size. If should be 4 for normal copy,
8 or 12 for some conditional branch. */
if ((*adjusted_insn_addr_end - *adjusted_insn_addr == 0)
|| (*adjusted_insn_addr_end - *adjusted_insn_addr > 12))
{
sprintf (err, "E.Unexpected instruction length = %d"
"when relocate instruction.",
(int) (*adjusted_insn_addr_end - *adjusted_insn_addr));
return 1;
}
buildaddr = *adjusted_insn_addr_end;
p = buf;
/* Finally, write a jump back to the program. */
offset = (tpaddr + 4) - buildaddr;
if (offset >= (1 << 25) || offset < -(1 << 25))
{
sprintf (err, "E.Jump back from jump pad too far from tracepoint "
"(offset 0x%x > 26-bit).", offset);
return 1;
}
/* b <tpaddr+4> */
p += GEN_B (p, offset);
write_inferior_memory (buildaddr, (unsigned char *) buf, (p - buf) * 4);
*jump_entry = buildaddr + (p - buf) * 4;
/* The jump pad is now built. Wire in a jump to our jump pad. This
is always done last (by our caller actually), so that we can
install fast tracepoints with threads running. This relies on
the agent's atomic write support. */
offset = entryaddr - tpaddr;
if (offset >= (1 << 25) || offset < -(1 << 25))
{
sprintf (err, "E.Jump back from jump pad too far from tracepoint "
"(offset 0x%x > 26-bit).", offset);
return 1;
}
/* b <jentry> */
GEN_B ((uint32_t *) jjump_pad_insn, offset);
*jjump_pad_insn_size = 4;
return 0;
}
/* Returns the minimum instruction length for installing a tracepoint. */
static int
ppc_get_min_fast_tracepoint_insn_len (void)
{
return 4;
}
/* Emits a given buffer into the target at current_insn_ptr. Length
is in units of 32-bit words. */
static void
emit_insns (uint32_t *buf, int n)
{
n = n * sizeof (uint32_t);
write_inferior_memory (current_insn_ptr, (unsigned char *) buf, n);
current_insn_ptr += n;
}
#define __EMIT_ASM(NAME, INSNS) \
do \
{ \
extern uint32_t start_bcax_ ## NAME []; \
extern uint32_t end_bcax_ ## NAME []; \
emit_insns (start_bcax_ ## NAME, \
end_bcax_ ## NAME - start_bcax_ ## NAME); \
__asm__ (".section .text.__ppcbcax\n\t" \
"start_bcax_" #NAME ":\n\t" \
INSNS "\n\t" \
"end_bcax_" #NAME ":\n\t" \
".previous\n\t"); \
} while (0)
#define _EMIT_ASM(NAME, INSNS) __EMIT_ASM (NAME, INSNS)
#define EMIT_ASM(INSNS) _EMIT_ASM (__LINE__, INSNS)
/*
Bytecode execution stack frame - 32-bit
| LR save area (SP + 4)
SP' -> +- Back chain (SP + 0)
| Save r31 for access saved arguments
| Save r30 for bytecode stack pointer
| Save r4 for incoming argument *value
| Save r3 for incoming argument regs
r30 -> +- Bytecode execution stack
|
| 64-byte (8 doublewords) at initial.
| Expand stack as needed.
|
+-
| Some padding for minimum stack frame and 16-byte alignment.
| 16 bytes.
SP +- Back-chain (SP')
initial frame size
= 16 + (4 * 4) + 64
= 96
r30 is the stack-pointer for bytecode machine.
It should point to next-empty, so we can use LDU for pop.
r3 is used for cache of the high part of TOP value.
It was the first argument, pointer to regs.
r4 is used for cache of the low part of TOP value.
It was the second argument, pointer to the result.
We should set *result = TOP after leaving this function.
Note:
* To restore stack at epilogue
=> sp = r31
* To check stack is big enough for bytecode execution.
=> r30 - 8 > SP + 8
* To return execution result.
=> 0(r4) = TOP
*/
/* Regardless of endian, register 3 is always high part, 4 is low part.
These defines are used when the register pair is stored/loaded.
Likewise, to simplify code, have a similiar define for 5:6. */
#if __BYTE_ORDER == __LITTLE_ENDIAN
#define TOP_FIRST "4"
#define TOP_SECOND "3"
#define TMP_FIRST "6"
#define TMP_SECOND "5"
#else
#define TOP_FIRST "3"
#define TOP_SECOND "4"
#define TMP_FIRST "5"
#define TMP_SECOND "6"
#endif
/* Emit prologue in inferior memory. See above comments. */
static void
ppc_emit_prologue (void)
{
EMIT_ASM (/* Save return address. */
"mflr 0 \n"
"stw 0, 4(1) \n"
/* Adjust SP. 96 is the initial frame size. */
"stwu 1, -96(1) \n"
/* Save r30 and incoming arguments. */
"stw 31, 96-4(1) \n"
"stw 30, 96-8(1) \n"
"stw 4, 96-12(1) \n"
"stw 3, 96-16(1) \n"
/* Point r31 to original r1 for access arguments. */
"addi 31, 1, 96 \n"
/* Set r30 to pointing stack-top. */
"addi 30, 1, 64 \n"
/* Initial r3/TOP to 0. */
"li 3, 0 \n"
"li 4, 0 \n");
}
/* Emit epilogue in inferior memory. See above comments. */
static void
ppc_emit_epilogue (void)
{
EMIT_ASM (/* *result = TOP */
"lwz 5, -12(31) \n"
"stw " TOP_FIRST ", 0(5) \n"
"stw " TOP_SECOND ", 4(5) \n"
/* Restore registers. */
"lwz 31, -4(31) \n"
"lwz 30, -8(31) \n"
/* Restore SP. */
"lwz 1, 0(1) \n"
/* Restore LR. */
"lwz 0, 4(1) \n"
/* Return 0 for no-error. */
"li 3, 0 \n"
"mtlr 0 \n"
"blr \n");
}
/* TOP = stack[--sp] + TOP */
static void
ppc_emit_add (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30)\n"
"addc 4, 6, 4 \n"
"adde 3, 5, 3 \n");
}
/* TOP = stack[--sp] - TOP */
static void
ppc_emit_sub (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"subfc 4, 4, 6 \n"
"subfe 3, 3, 5 \n");
}
/* TOP = stack[--sp] * TOP */
static void
ppc_emit_mul (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"mulhwu 7, 6, 4 \n"
"mullw 3, 6, 3 \n"
"mullw 5, 4, 5 \n"
"mullw 4, 6, 4 \n"
"add 3, 5, 3 \n"
"add 3, 7, 3 \n");
}
/* TOP = stack[--sp] << TOP */
static void
ppc_emit_lsh (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"subfic 3, 4, 32\n" /* r3 = 32 - TOP */
"addi 7, 4, -32\n" /* r7 = TOP - 32 */
"slw 5, 5, 4\n" /* Shift high part left */
"slw 4, 6, 4\n" /* Shift low part left */
"srw 3, 6, 3\n" /* Shift low to high if shift < 32 */
"slw 7, 6, 7\n" /* Shift low to high if shift >= 32 */
"or 3, 5, 3\n"
"or 3, 7, 3\n"); /* Assemble high part */
}
/* Top = stack[--sp] >> TOP
(Arithmetic shift right) */
static void
ppc_emit_rsh_signed (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"addi 7, 4, -32\n" /* r7 = TOP - 32 */
"sraw 3, 5, 4\n" /* Shift high part right */
"cmpwi 7, 1\n"
"blt 0, 1f\n" /* If shift <= 32, goto 1: */
"sraw 4, 5, 7\n" /* Shift high to low */
"b 2f\n"
"1:\n"
"subfic 7, 4, 32\n" /* r7 = 32 - TOP */
"srw 4, 6, 4\n" /* Shift low part right */
"slw 5, 5, 7\n" /* Shift high to low */
"or 4, 4, 5\n" /* Assemble low part */
"2:\n");
}
/* Top = stack[--sp] >> TOP
(Logical shift right) */
static void
ppc_emit_rsh_unsigned (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"subfic 3, 4, 32\n" /* r3 = 32 - TOP */
"addi 7, 4, -32\n" /* r7 = TOP - 32 */
"srw 6, 6, 4\n" /* Shift low part right */
"slw 3, 5, 3\n" /* Shift high to low if shift < 32 */
"srw 7, 5, 7\n" /* Shift high to low if shift >= 32 */
"or 6, 6, 3\n"
"srw 3, 5, 4\n" /* Shift high part right */
"or 4, 6, 7\n"); /* Assemble low part */
}
/* Emit code for signed-extension specified by ARG. */
static void
ppc_emit_ext (int arg)
{
switch (arg)
{
case 8:
EMIT_ASM ("extsb 4, 4\n"
"srawi 3, 4, 31");
break;
case 16:
EMIT_ASM ("extsh 4, 4\n"
"srawi 3, 4, 31");
break;
case 32:
EMIT_ASM ("srawi 3, 4, 31");
break;
default:
emit_error = 1;
}
}
/* Emit code for zero-extension specified by ARG. */
static void
ppc_emit_zero_ext (int arg)
{
switch (arg)
{
case 8:
EMIT_ASM ("clrlwi 4,4,24\n"
"li 3, 0\n");
break;
case 16:
EMIT_ASM ("clrlwi 4,4,16\n"
"li 3, 0\n");
break;
case 32:
EMIT_ASM ("li 3, 0");
break;
default:
emit_error = 1;
}
}
/* TOP = !TOP
i.e., TOP = (TOP == 0) ? 1 : 0; */
static void
ppc_emit_log_not (void)
{
EMIT_ASM ("or 4, 3, 4 \n"
"cntlzw 4, 4 \n"
"srwi 4, 4, 5 \n"
"li 3, 0 \n");
}
/* TOP = stack[--sp] & TOP */
static void
ppc_emit_bit_and (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"and 4, 6, 4 \n"
"and 3, 5, 3 \n");
}
/* TOP = stack[--sp] | TOP */
static void
ppc_emit_bit_or (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"or 4, 6, 4 \n"
"or 3, 5, 3 \n");
}
/* TOP = stack[--sp] ^ TOP */
static void
ppc_emit_bit_xor (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"xor 4, 6, 4 \n"
"xor 3, 5, 3 \n");
}
/* TOP = ~TOP
i.e., TOP = ~(TOP | TOP) */
static void
ppc_emit_bit_not (void)
{
EMIT_ASM ("nor 3, 3, 3 \n"
"nor 4, 4, 4 \n");
}
/* TOP = stack[--sp] == TOP */
static void
ppc_emit_equal (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"xor 4, 6, 4 \n"
"xor 3, 5, 3 \n"
"or 4, 3, 4 \n"
"cntlzw 4, 4 \n"
"srwi 4, 4, 5 \n"
"li 3, 0 \n");
}
/* TOP = stack[--sp] < TOP
(Signed comparison) */
static void
ppc_emit_less_signed (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmpw 7, 5, 3 \n"
/* CR6 bit 0 = low less and high equal */
"crand 6*4+0, 6*4+0, 7*4+2\n"
/* CR7 bit 0 = (low less and high equal) or high less */
"cror 7*4+0, 7*4+0, 6*4+0\n"
"mfcr 4 \n"
"rlwinm 4, 4, 29, 31, 31 \n"
"li 3, 0 \n");
}
/* TOP = stack[--sp] < TOP
(Unsigned comparison) */
static void
ppc_emit_less_unsigned (void)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmplw 7, 5, 3 \n"
/* CR6 bit 0 = low less and high equal */
"crand 6*4+0, 6*4+0, 7*4+2\n"
/* CR7 bit 0 = (low less and high equal) or high less */
"cror 7*4+0, 7*4+0, 6*4+0\n"
"mfcr 4 \n"
"rlwinm 4, 4, 29, 31, 31 \n"
"li 3, 0 \n");
}
/* Access the memory address in TOP in size of SIZE.
Zero-extend the read value. */
static void
ppc_emit_ref (int size)
{
switch (size)
{
case 1:
EMIT_ASM ("lbz 4, 0(4)\n"
"li 3, 0");
break;
case 2:
EMIT_ASM ("lhz 4, 0(4)\n"
"li 3, 0");
break;
case 4:
EMIT_ASM ("lwz 4, 0(4)\n"
"li 3, 0");
break;
case 8:
if (__BYTE_ORDER == __LITTLE_ENDIAN)
EMIT_ASM ("lwz 3, 4(4)\n"
"lwz 4, 0(4)");
else
EMIT_ASM ("lwz 3, 0(4)\n"
"lwz 4, 4(4)");
break;
}
}
/* TOP = NUM */
static void
ppc_emit_const (LONGEST num)
{
uint32_t buf[10];
uint32_t *p = buf;
p += gen_limm (p, 3, num >> 32 & 0xffffffff, 0);
p += gen_limm (p, 4, num & 0xffffffff, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Set TOP to the value of register REG by calling get_raw_reg function
with two argument, collected buffer and register number. */
static void
ppc_emit_reg (int reg)
{
uint32_t buf[13];
uint32_t *p = buf;
/* fctx->regs is passed in r3 and then saved in -16(31). */
p += GEN_LWZ (p, 3, 31, -16);
p += GEN_LI (p, 4, reg); /* li r4, reg */
p += gen_call (p, get_raw_reg_func_addr (), 0, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
if (__BYTE_ORDER == __LITTLE_ENDIAN)
{
EMIT_ASM ("mr 5, 4\n"
"mr 4, 3\n"
"mr 3, 5\n");
}
}
/* TOP = stack[--sp] */
static void
ppc_emit_pop (void)
{
EMIT_ASM ("lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30) \n");
}
/* stack[sp++] = TOP
Because we may use up bytecode stack, expand 8 doublewords more
if needed. */
static void
ppc_emit_stack_flush (void)
{
/* Make sure bytecode stack is big enough before push.
Otherwise, expand 64-byte more. */
EMIT_ASM (" stw " TOP_FIRST ", 0(30) \n"
" stw " TOP_SECOND ", 4(30)\n"
" addi 5, 30, -(8 + 8) \n"
" cmpw 7, 5, 1 \n"
" bgt 7, 1f \n"
" stwu 31, -64(1) \n"
"1:addi 30, 30, -8 \n");
}
/* Swap TOP and stack[sp-1] */
static void
ppc_emit_swap (void)
{
EMIT_ASM ("lwz " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 12(30) \n"
"stw " TOP_FIRST ", 8(30) \n"
"stw " TOP_SECOND ", 12(30) \n"
"mr 3, 5 \n"
"mr 4, 6 \n");
}
/* Discard N elements in the stack. Also used for ppc64. */
static void
ppc_emit_stack_adjust (int n)
{
uint32_t buf[6];
uint32_t *p = buf;
n = n << 3;
if ((n >> 15) != 0)
{
emit_error = 1;
return;
}
p += GEN_ADDI (p, 30, 30, n);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Call function FN. */
static void
ppc_emit_call (CORE_ADDR fn)
{
uint32_t buf[11];
uint32_t *p = buf;
p += gen_call (p, fn, 0, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* FN's prototype is `LONGEST(*fn)(int)'.
TOP = fn (arg1)
*/
static void
ppc_emit_int_call_1 (CORE_ADDR fn, int arg1)
{
uint32_t buf[15];
uint32_t *p = buf;
/* Setup argument. arg1 is a 16-bit value. */
p += gen_limm (p, 3, (uint32_t) arg1, 0);
p += gen_call (p, fn, 0, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
if (__BYTE_ORDER == __LITTLE_ENDIAN)
{
EMIT_ASM ("mr 5, 4\n"
"mr 4, 3\n"
"mr 3, 5\n");
}
}
/* FN's prototype is `void(*fn)(int,LONGEST)'.
fn (arg1, TOP)
TOP should be preserved/restored before/after the call. */
static void
ppc_emit_void_call_2 (CORE_ADDR fn, int arg1)
{
uint32_t buf[21];
uint32_t *p = buf;
/* Save TOP. 0(30) is next-empty. */
p += GEN_STW (p, 3, 30, 0);
p += GEN_STW (p, 4, 30, 4);
/* Setup argument. arg1 is a 16-bit value. */
if (__BYTE_ORDER == __LITTLE_ENDIAN)
{
p += GEN_MR (p, 5, 4);
p += GEN_MR (p, 6, 3);
}
else
{
p += GEN_MR (p, 5, 3);
p += GEN_MR (p, 6, 4);
}
p += gen_limm (p, 3, (uint32_t) arg1, 0);
p += gen_call (p, fn, 0, 0);
/* Restore TOP */
p += GEN_LWZ (p, 3, 30, 0);
p += GEN_LWZ (p, 4, 30, 4);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Note in the following goto ops:
When emitting goto, the target address is later relocated by
write_goto_address. OFFSET_P is the offset of the branch instruction
in the code sequence, and SIZE_P is how to relocate the instruction,
recognized by ppc_write_goto_address. In current implementation,
SIZE can be either 24 or 14 for branch of conditional-branch instruction.
*/
/* If TOP is true, goto somewhere. Otherwise, just fall-through. */
static void
ppc_emit_if_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("or. 3, 3, 4 \n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30) \n"
"1:bne 0, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Unconditional goto. Also used for ppc64. */
static void
ppc_emit_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("1:b 1b");
if (offset_p)
*offset_p = 0;
if (size_p)
*size_p = 24;
}
/* Goto if stack[--sp] == TOP */
static void
ppc_emit_eq_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"xor 4, 6, 4 \n"
"xor 3, 5, 3 \n"
"or. 3, 3, 4 \n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30) \n"
"1:beq 0, 1b \n");
if (offset_p)
*offset_p = 28;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] != TOP */
static void
ppc_emit_ne_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"xor 4, 6, 4 \n"
"xor 3, 5, 3 \n"
"or. 3, 3, 4 \n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30) \n"
"1:bne 0, 1b \n");
if (offset_p)
*offset_p = 28;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] < TOP */
static void
ppc_emit_lt_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmpw 7, 5, 3 \n"
/* CR6 bit 0 = low less and high equal */
"crand 6*4+0, 6*4+0, 7*4+2\n"
/* CR7 bit 0 = (low less and high equal) or high less */
"cror 7*4+0, 7*4+0, 6*4+0\n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30)\n"
"1:blt 7, 1b \n");
if (offset_p)
*offset_p = 32;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] <= TOP */
static void
ppc_emit_le_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmpw 7, 5, 3 \n"
/* CR6 bit 0 = low less/equal and high equal */
"crandc 6*4+0, 7*4+2, 6*4+1\n"
/* CR7 bit 0 = (low less/eq and high equal) or high less */
"cror 7*4+0, 7*4+0, 6*4+0\n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30)\n"
"1:blt 7, 1b \n");
if (offset_p)
*offset_p = 32;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] > TOP */
static void
ppc_emit_gt_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmpw 7, 5, 3 \n"
/* CR6 bit 0 = low greater and high equal */
"crand 6*4+0, 6*4+1, 7*4+2\n"
/* CR7 bit 0 = (low greater and high equal) or high greater */
"cror 7*4+0, 7*4+1, 6*4+0\n"
"lwzu " TOP_FIRST ", 8(30) \n"
"lwz " TOP_SECOND ", 4(30)\n"
"1:blt 7, 1b \n");
if (offset_p)
*offset_p = 32;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] >= TOP */
static void
ppc_emit_ge_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("lwzu " TMP_FIRST ", 8(30) \n"
"lwz " TMP_SECOND ", 4(30) \n"
"cmplw 6, 6, 4 \n"
"cmpw 7, 5, 3 \n"
/* CR6 bit 0 = low ge and high equal */
"crandc 6*4+0, 7*4+2, 6*4+0\n"
/* CR7 bit 0 = (low ge and high equal) or high greater */
"cror 7*4+0, 7*4+1, 6*4+0\n"
"lwzu " TOP_FIRST ", 8(30)\n"
"lwz " TOP_SECOND ", 4(30)\n"
"1:blt 7, 1b \n");
if (offset_p)
*offset_p = 32;
if (size_p)
*size_p = 14;
}
/* Relocate previous emitted branch instruction. FROM is the address
of the branch instruction, TO is the goto target address, and SIZE
if the value we set by *SIZE_P before. Currently, it is either
24 or 14 of branch and conditional-branch instruction.
Also used for ppc64. */
static void
ppc_write_goto_address (CORE_ADDR from, CORE_ADDR to, int size)
{
long rel = to - from;
uint32_t insn;
int opcd;
read_inferior_memory (from, (unsigned char *) &insn, 4);
opcd = (insn >> 26) & 0x3f;
switch (size)
{
case 14:
if (opcd != 16
|| (rel >= (1 << 15) || rel < -(1 << 15)))
emit_error = 1;
insn = (insn & ~0xfffc) | (rel & 0xfffc);
break;
case 24:
if (opcd != 18
|| (rel >= (1 << 25) || rel < -(1 << 25)))
emit_error = 1;
insn = (insn & ~0x3fffffc) | (rel & 0x3fffffc);
break;
default:
emit_error = 1;
}
if (!emit_error)
write_inferior_memory (from, (unsigned char *) &insn, 4);
}
/* Table of emit ops for 32-bit. */
static struct emit_ops ppc_emit_ops_impl =
{
ppc_emit_prologue,
ppc_emit_epilogue,
ppc_emit_add,
ppc_emit_sub,
ppc_emit_mul,
ppc_emit_lsh,
ppc_emit_rsh_signed,
ppc_emit_rsh_unsigned,
ppc_emit_ext,
ppc_emit_log_not,
ppc_emit_bit_and,
ppc_emit_bit_or,
ppc_emit_bit_xor,
ppc_emit_bit_not,
ppc_emit_equal,
ppc_emit_less_signed,
ppc_emit_less_unsigned,
ppc_emit_ref,
ppc_emit_if_goto,
ppc_emit_goto,
ppc_write_goto_address,
ppc_emit_const,
ppc_emit_call,
ppc_emit_reg,
ppc_emit_pop,
ppc_emit_stack_flush,
ppc_emit_zero_ext,
ppc_emit_swap,
ppc_emit_stack_adjust,
ppc_emit_int_call_1,
ppc_emit_void_call_2,
ppc_emit_eq_goto,
ppc_emit_ne_goto,
ppc_emit_lt_goto,
ppc_emit_le_goto,
ppc_emit_gt_goto,
ppc_emit_ge_goto
};
#ifdef __powerpc64__
/*
Bytecode execution stack frame - 64-bit
| LR save area (SP + 16)
| CR save area (SP + 8)
SP' -> +- Back chain (SP + 0)
| Save r31 for access saved arguments
| Save r30 for bytecode stack pointer
| Save r4 for incoming argument *value
| Save r3 for incoming argument regs
r30 -> +- Bytecode execution stack
|
| 64-byte (8 doublewords) at initial.
| Expand stack as needed.
|
+-
| Some padding for minimum stack frame.
| 112 for ELFv1.
SP +- Back-chain (SP')
initial frame size
= 112 + (4 * 8) + 64
= 208
r30 is the stack-pointer for bytecode machine.
It should point to next-empty, so we can use LDU for pop.
r3 is used for cache of TOP value.
It was the first argument, pointer to regs.
r4 is the second argument, pointer to the result.
We should set *result = TOP after leaving this function.
Note:
* To restore stack at epilogue
=> sp = r31
* To check stack is big enough for bytecode execution.
=> r30 - 8 > SP + 112
* To return execution result.
=> 0(r4) = TOP
*/
/* Emit prologue in inferior memory. See above comments. */
static void
ppc64v1_emit_prologue (void)
{
/* On ELFv1, function pointers really point to function descriptor,
so emit one here. We don't care about contents of words 1 and 2,
so let them just overlap out code. */
uint64_t opd = current_insn_ptr + 8;
uint32_t buf[2];
/* Mind the strict aliasing rules. */
memcpy (buf, &opd, sizeof buf);
emit_insns(buf, 2);
EMIT_ASM (/* Save return address. */
"mflr 0 \n"
"std 0, 16(1) \n"
/* Save r30 and incoming arguments. */
"std 31, -8(1) \n"
"std 30, -16(1) \n"
"std 4, -24(1) \n"
"std 3, -32(1) \n"
/* Point r31 to current r1 for access arguments. */
"mr 31, 1 \n"
/* Adjust SP. 208 is the initial frame size. */
"stdu 1, -208(1) \n"
/* Set r30 to pointing stack-top. */
"addi 30, 1, 168 \n"
/* Initial r3/TOP to 0. */
"li 3, 0 \n");
}
/* Emit prologue in inferior memory. See above comments. */
static void
ppc64v2_emit_prologue (void)
{
EMIT_ASM (/* Save return address. */
"mflr 0 \n"
"std 0, 16(1) \n"
/* Save r30 and incoming arguments. */
"std 31, -8(1) \n"
"std 30, -16(1) \n"
"std 4, -24(1) \n"
"std 3, -32(1) \n"
/* Point r31 to current r1 for access arguments. */
"mr 31, 1 \n"
/* Adjust SP. 208 is the initial frame size. */
"stdu 1, -208(1) \n"
/* Set r30 to pointing stack-top. */
"addi 30, 1, 168 \n"
/* Initial r3/TOP to 0. */
"li 3, 0 \n");
}
/* Emit epilogue in inferior memory. See above comments. */
static void
ppc64_emit_epilogue (void)
{
EMIT_ASM (/* Restore SP. */
"ld 1, 0(1) \n"
/* *result = TOP */
"ld 4, -24(1) \n"
"std 3, 0(4) \n"
/* Restore registers. */
"ld 31, -8(1) \n"
"ld 30, -16(1) \n"
/* Restore LR. */
"ld 0, 16(1) \n"
/* Return 0 for no-error. */
"li 3, 0 \n"
"mtlr 0 \n"
"blr \n");
}
/* TOP = stack[--sp] + TOP */
static void
ppc64_emit_add (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"add 3, 4, 3 \n");
}
/* TOP = stack[--sp] - TOP */
static void
ppc64_emit_sub (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"sub 3, 4, 3 \n");
}
/* TOP = stack[--sp] * TOP */
static void
ppc64_emit_mul (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"mulld 3, 4, 3 \n");
}
/* TOP = stack[--sp] << TOP */
static void
ppc64_emit_lsh (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"sld 3, 4, 3 \n");
}
/* Top = stack[--sp] >> TOP
(Arithmetic shift right) */
static void
ppc64_emit_rsh_signed (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"srad 3, 4, 3 \n");
}
/* Top = stack[--sp] >> TOP
(Logical shift right) */
static void
ppc64_emit_rsh_unsigned (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"srd 3, 4, 3 \n");
}
/* Emit code for signed-extension specified by ARG. */
static void
ppc64_emit_ext (int arg)
{
switch (arg)
{
case 8:
EMIT_ASM ("extsb 3, 3");
break;
case 16:
EMIT_ASM ("extsh 3, 3");
break;
case 32:
EMIT_ASM ("extsw 3, 3");
break;
default:
emit_error = 1;
}
}
/* Emit code for zero-extension specified by ARG. */
static void
ppc64_emit_zero_ext (int arg)
{
switch (arg)
{
case 8:
EMIT_ASM ("rldicl 3,3,0,56");
break;
case 16:
EMIT_ASM ("rldicl 3,3,0,48");
break;
case 32:
EMIT_ASM ("rldicl 3,3,0,32");
break;
default:
emit_error = 1;
}
}
/* TOP = !TOP
i.e., TOP = (TOP == 0) ? 1 : 0; */
static void
ppc64_emit_log_not (void)
{
EMIT_ASM ("cntlzd 3, 3 \n"
"srdi 3, 3, 6 \n");
}
/* TOP = stack[--sp] & TOP */
static void
ppc64_emit_bit_and (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"and 3, 4, 3 \n");
}
/* TOP = stack[--sp] | TOP */
static void
ppc64_emit_bit_or (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"or 3, 4, 3 \n");
}
/* TOP = stack[--sp] ^ TOP */
static void
ppc64_emit_bit_xor (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"xor 3, 4, 3 \n");
}
/* TOP = ~TOP
i.e., TOP = ~(TOP | TOP) */
static void
ppc64_emit_bit_not (void)
{
EMIT_ASM ("nor 3, 3, 3 \n");
}
/* TOP = stack[--sp] == TOP */
static void
ppc64_emit_equal (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"xor 3, 3, 4 \n"
"cntlzd 3, 3 \n"
"srdi 3, 3, 6 \n");
}
/* TOP = stack[--sp] < TOP
(Signed comparison) */
static void
ppc64_emit_less_signed (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"mfcr 3 \n"
"rlwinm 3, 3, 29, 31, 31 \n");
}
/* TOP = stack[--sp] < TOP
(Unsigned comparison) */
static void
ppc64_emit_less_unsigned (void)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpld 7, 4, 3 \n"
"mfcr 3 \n"
"rlwinm 3, 3, 29, 31, 31 \n");
}
/* Access the memory address in TOP in size of SIZE.
Zero-extend the read value. */
static void
ppc64_emit_ref (int size)
{
switch (size)
{
case 1:
EMIT_ASM ("lbz 3, 0(3)");
break;
case 2:
EMIT_ASM ("lhz 3, 0(3)");
break;
case 4:
EMIT_ASM ("lwz 3, 0(3)");
break;
case 8:
EMIT_ASM ("ld 3, 0(3)");
break;
}
}
/* TOP = NUM */
static void
ppc64_emit_const (LONGEST num)
{
uint32_t buf[5];
uint32_t *p = buf;
p += gen_limm (p, 3, num, 1);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Set TOP to the value of register REG by calling get_raw_reg function
with two argument, collected buffer and register number. */
static void
ppc64v1_emit_reg (int reg)
{
uint32_t buf[15];
uint32_t *p = buf;
/* fctx->regs is passed in r3 and then saved in 176(1). */
p += GEN_LD (p, 3, 31, -32);
p += GEN_LI (p, 4, reg);
p += GEN_STD (p, 2, 1, 40); /* Save TOC. */
p += gen_call (p, get_raw_reg_func_addr (), 1, 1);
p += GEN_LD (p, 2, 1, 40); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Likewise, for ELFv2. */
static void
ppc64v2_emit_reg (int reg)
{
uint32_t buf[12];
uint32_t *p = buf;
/* fctx->regs is passed in r3 and then saved in 176(1). */
p += GEN_LD (p, 3, 31, -32);
p += GEN_LI (p, 4, reg);
p += GEN_STD (p, 2, 1, 24); /* Save TOC. */
p += gen_call (p, get_raw_reg_func_addr (), 1, 0);
p += GEN_LD (p, 2, 1, 24); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* TOP = stack[--sp] */
static void
ppc64_emit_pop (void)
{
EMIT_ASM ("ldu 3, 8(30)");
}
/* stack[sp++] = TOP
Because we may use up bytecode stack, expand 8 doublewords more
if needed. */
static void
ppc64_emit_stack_flush (void)
{
/* Make sure bytecode stack is big enough before push.
Otherwise, expand 64-byte more. */
EMIT_ASM (" std 3, 0(30) \n"
" addi 4, 30, -(112 + 8) \n"
" cmpd 7, 4, 1 \n"
" bgt 7, 1f \n"
" stdu 31, -64(1) \n"
"1:addi 30, 30, -8 \n");
}
/* Swap TOP and stack[sp-1] */
static void
ppc64_emit_swap (void)
{
EMIT_ASM ("ld 4, 8(30) \n"
"std 3, 8(30) \n"
"mr 3, 4 \n");
}
/* Call function FN - ELFv1. */
static void
ppc64v1_emit_call (CORE_ADDR fn)
{
uint32_t buf[13];
uint32_t *p = buf;
p += GEN_STD (p, 2, 1, 40); /* Save TOC. */
p += gen_call (p, fn, 1, 1);
p += GEN_LD (p, 2, 1, 40); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Call function FN - ELFv2. */
static void
ppc64v2_emit_call (CORE_ADDR fn)
{
uint32_t buf[10];
uint32_t *p = buf;
p += GEN_STD (p, 2, 1, 24); /* Save TOC. */
p += gen_call (p, fn, 1, 0);
p += GEN_LD (p, 2, 1, 24); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* FN's prototype is `LONGEST(*fn)(int)'.
TOP = fn (arg1)
*/
static void
ppc64v1_emit_int_call_1 (CORE_ADDR fn, int arg1)
{
uint32_t buf[13];
uint32_t *p = buf;
/* Setup argument. arg1 is a 16-bit value. */
p += gen_limm (p, 3, arg1, 1);
p += GEN_STD (p, 2, 1, 40); /* Save TOC. */
p += gen_call (p, fn, 1, 1);
p += GEN_LD (p, 2, 1, 40); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Likewise for ELFv2. */
static void
ppc64v2_emit_int_call_1 (CORE_ADDR fn, int arg1)
{
uint32_t buf[10];
uint32_t *p = buf;
/* Setup argument. arg1 is a 16-bit value. */
p += gen_limm (p, 3, arg1, 1);
p += GEN_STD (p, 2, 1, 24); /* Save TOC. */
p += gen_call (p, fn, 1, 0);
p += GEN_LD (p, 2, 1, 24); /* Restore TOC. */
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* FN's prototype is `void(*fn)(int,LONGEST)'.
fn (arg1, TOP)
TOP should be preserved/restored before/after the call. */
static void
ppc64v1_emit_void_call_2 (CORE_ADDR fn, int arg1)
{
uint32_t buf[17];
uint32_t *p = buf;
/* Save TOP. 0(30) is next-empty. */
p += GEN_STD (p, 3, 30, 0);
/* Setup argument. arg1 is a 16-bit value. */
p += GEN_MR (p, 4, 3); /* mr r4, r3 */
p += gen_limm (p, 3, arg1, 1);
p += GEN_STD (p, 2, 1, 40); /* Save TOC. */
p += gen_call (p, fn, 1, 1);
p += GEN_LD (p, 2, 1, 40); /* Restore TOC. */
/* Restore TOP */
p += GEN_LD (p, 3, 30, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* Likewise for ELFv2. */
static void
ppc64v2_emit_void_call_2 (CORE_ADDR fn, int arg1)
{
uint32_t buf[14];
uint32_t *p = buf;
/* Save TOP. 0(30) is next-empty. */
p += GEN_STD (p, 3, 30, 0);
/* Setup argument. arg1 is a 16-bit value. */
p += GEN_MR (p, 4, 3); /* mr r4, r3 */
p += gen_limm (p, 3, arg1, 1);
p += GEN_STD (p, 2, 1, 24); /* Save TOC. */
p += gen_call (p, fn, 1, 0);
p += GEN_LD (p, 2, 1, 24); /* Restore TOC. */
/* Restore TOP */
p += GEN_LD (p, 3, 30, 0);
emit_insns (buf, p - buf);
gdb_assert ((p - buf) <= (sizeof (buf) / sizeof (*buf)));
}
/* If TOP is true, goto somewhere. Otherwise, just fall-through. */
static void
ppc64_emit_if_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("cmpdi 7, 3, 0 \n"
"ldu 3, 8(30) \n"
"1:bne 7, 1b \n");
if (offset_p)
*offset_p = 8;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] == TOP */
static void
ppc64_emit_eq_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:beq 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] != TOP */
static void
ppc64_emit_ne_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:bne 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] < TOP */
static void
ppc64_emit_lt_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:blt 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] <= TOP */
static void
ppc64_emit_le_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:ble 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] > TOP */
static void
ppc64_emit_gt_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:bgt 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Goto if stack[--sp] >= TOP */
static void
ppc64_emit_ge_goto (int *offset_p, int *size_p)
{
EMIT_ASM ("ldu 4, 8(30) \n"
"cmpd 7, 4, 3 \n"
"ldu 3, 8(30) \n"
"1:bge 7, 1b \n");
if (offset_p)
*offset_p = 12;
if (size_p)
*size_p = 14;
}
/* Table of emit ops for 64-bit ELFv1. */
static struct emit_ops ppc64v1_emit_ops_impl =
{
ppc64v1_emit_prologue,
ppc64_emit_epilogue,
ppc64_emit_add,
ppc64_emit_sub,
ppc64_emit_mul,
ppc64_emit_lsh,
ppc64_emit_rsh_signed,
ppc64_emit_rsh_unsigned,
ppc64_emit_ext,
ppc64_emit_log_not,
ppc64_emit_bit_and,
ppc64_emit_bit_or,
ppc64_emit_bit_xor,
ppc64_emit_bit_not,
ppc64_emit_equal,
ppc64_emit_less_signed,
ppc64_emit_less_unsigned,
ppc64_emit_ref,
ppc64_emit_if_goto,
ppc_emit_goto,
ppc_write_goto_address,
ppc64_emit_const,
ppc64v1_emit_call,
ppc64v1_emit_reg,
ppc64_emit_pop,
ppc64_emit_stack_flush,
ppc64_emit_zero_ext,
ppc64_emit_swap,
ppc_emit_stack_adjust,
ppc64v1_emit_int_call_1,
ppc64v1_emit_void_call_2,
ppc64_emit_eq_goto,
ppc64_emit_ne_goto,
ppc64_emit_lt_goto,
ppc64_emit_le_goto,
ppc64_emit_gt_goto,
ppc64_emit_ge_goto
};
/* Table of emit ops for 64-bit ELFv2. */
static struct emit_ops ppc64v2_emit_ops_impl =
{
ppc64v2_emit_prologue,
ppc64_emit_epilogue,
ppc64_emit_add,
ppc64_emit_sub,
ppc64_emit_mul,
ppc64_emit_lsh,
ppc64_emit_rsh_signed,
ppc64_emit_rsh_unsigned,
ppc64_emit_ext,
ppc64_emit_log_not,
ppc64_emit_bit_and,
ppc64_emit_bit_or,
ppc64_emit_bit_xor,
ppc64_emit_bit_not,
ppc64_emit_equal,
ppc64_emit_less_signed,
ppc64_emit_less_unsigned,
ppc64_emit_ref,
ppc64_emit_if_goto,
ppc_emit_goto,
ppc_write_goto_address,
ppc64_emit_const,
ppc64v2_emit_call,
ppc64v2_emit_reg,
ppc64_emit_pop,
ppc64_emit_stack_flush,
ppc64_emit_zero_ext,
ppc64_emit_swap,
ppc_emit_stack_adjust,
ppc64v2_emit_int_call_1,
ppc64v2_emit_void_call_2,
ppc64_emit_eq_goto,
ppc64_emit_ne_goto,
ppc64_emit_lt_goto,
ppc64_emit_le_goto,
ppc64_emit_gt_goto,
ppc64_emit_ge_goto
};
#endif
/* Implementation of linux_target_ops method "emit_ops". */
static struct emit_ops *
ppc_emit_ops (void)
{
#ifdef __powerpc64__
struct regcache *regcache = get_thread_regcache (current_thread, 0);
if (register_size (regcache->tdesc, 0) == 8)
{
if (is_elfv2_inferior ())
return &ppc64v2_emit_ops_impl;
else
return &ppc64v1_emit_ops_impl;
}
#endif
return &ppc_emit_ops_impl;
}
/* Implementation of linux_target_ops method "get_ipa_tdesc_idx". */
static int
ppc_get_ipa_tdesc_idx (void)
{
struct regcache *regcache = get_thread_regcache (current_thread, 0);
const struct target_desc *tdesc = regcache->tdesc;
#ifdef __powerpc64__
if (tdesc == tdesc_powerpc_64l)
return PPC_TDESC_BASE;
if (tdesc == tdesc_powerpc_altivec64l)
return PPC_TDESC_ALTIVEC;
if (tdesc == tdesc_powerpc_cell64l)
return PPC_TDESC_CELL;
if (tdesc == tdesc_powerpc_vsx64l)
return PPC_TDESC_VSX;
if (tdesc == tdesc_powerpc_isa205_64l)
return PPC_TDESC_ISA205;
if (tdesc == tdesc_powerpc_isa205_altivec64l)
return PPC_TDESC_ISA205_ALTIVEC;
if (tdesc == tdesc_powerpc_isa205_vsx64l)
return PPC_TDESC_ISA205_VSX;
#endif
if (tdesc == tdesc_powerpc_32l)
return PPC_TDESC_BASE;
if (tdesc == tdesc_powerpc_altivec32l)
return PPC_TDESC_ALTIVEC;
if (tdesc == tdesc_powerpc_cell32l)
return PPC_TDESC_CELL;
if (tdesc == tdesc_powerpc_vsx32l)
return PPC_TDESC_VSX;
if (tdesc == tdesc_powerpc_isa205_32l)
return PPC_TDESC_ISA205;
if (tdesc == tdesc_powerpc_isa205_altivec32l)
return PPC_TDESC_ISA205_ALTIVEC;
if (tdesc == tdesc_powerpc_isa205_vsx32l)
return PPC_TDESC_ISA205_VSX;
if (tdesc == tdesc_powerpc_e500l)
return PPC_TDESC_E500;
return 0;
}
struct linux_target_ops the_low_target = {
ppc_arch_setup,
ppc_regs_info,
ppc_cannot_fetch_register,
ppc_cannot_store_register,
NULL, /* fetch_register */
ppc_get_pc,
ppc_set_pc,
NULL, /* breakpoint_kind_from_pc */
ppc_sw_breakpoint_from_kind,
NULL,
0,
ppc_breakpoint_at,
ppc_supports_z_point_type,
ppc_insert_point,
ppc_remove_point,
NULL,
NULL,
ppc_collect_ptrace_register,
ppc_supply_ptrace_register,
NULL, /* siginfo_fixup */
NULL, /* new_process */
NULL, /* delete_process */
NULL, /* new_thread */
NULL, /* delete_thread */
NULL, /* new_fork */
NULL, /* prepare_to_resume */
NULL, /* process_qsupported */
ppc_supports_tracepoints,
ppc_get_thread_area,
ppc_install_fast_tracepoint_jump_pad,
ppc_emit_ops,
ppc_get_min_fast_tracepoint_insn_len,
NULL, /* supports_range_stepping */
NULL, /* breakpoint_kind_from_current_state */
ppc_supports_hardware_single_step,
NULL, /* get_syscall_trapinfo */
ppc_get_ipa_tdesc_idx,
};
void
initialize_low_arch (void)
{
/* Initialize the Linux target descriptions. */
init_registers_powerpc_32l ();
init_registers_powerpc_altivec32l ();
init_registers_powerpc_cell32l ();
init_registers_powerpc_vsx32l ();
init_registers_powerpc_isa205_32l ();
init_registers_powerpc_isa205_altivec32l ();
init_registers_powerpc_isa205_vsx32l ();
init_registers_powerpc_e500l ();
#if __powerpc64__
init_registers_powerpc_64l ();
init_registers_powerpc_altivec64l ();
init_registers_powerpc_cell64l ();
init_registers_powerpc_vsx64l ();
init_registers_powerpc_isa205_64l ();
init_registers_powerpc_isa205_altivec64l ();
init_registers_powerpc_isa205_vsx64l ();
#endif
initialize_regsets_info (&ppc_regsets_info);
}