mirror of
https://sourceware.org/git/binutils-gdb.git
synced 2024-11-30 21:44:19 +08:00
eb8c5f3f23
* elf32-rx.c (elf32_rx_relax_section): Correct width check.
3544 lines
90 KiB
C
3544 lines
90 KiB
C
/* Renesas RX specific support for 32-bit ELF.
|
||
Copyright (C) 2008, 2009, 2010, 2011
|
||
Free Software Foundation, Inc.
|
||
|
||
This file is part of BFD, the Binary File Descriptor library.
|
||
|
||
This program is free software; you can redistribute it and/or modify
|
||
it under the terms of the GNU General Public License as published by
|
||
the Free Software Foundation; either version 3 of the License, or
|
||
(at your option) any later version.
|
||
|
||
This program is distributed in the hope that it will be useful,
|
||
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
GNU General Public License for more details.
|
||
|
||
You should have received a copy of the GNU General Public License
|
||
along with this program; if not, write to the Free Software
|
||
Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */
|
||
|
||
#include "sysdep.h"
|
||
#include "bfd.h"
|
||
#include "bfd_stdint.h"
|
||
#include "libbfd.h"
|
||
#include "elf-bfd.h"
|
||
#include "elf/rx.h"
|
||
#include "libiberty.h"
|
||
|
||
#define RX_OPCODE_BIG_ENDIAN 0
|
||
|
||
/* This is a meta-target that's used only with objcopy, to avoid the
|
||
endian-swap we would otherwise get. We check for this in
|
||
rx_elf_object_p(). */
|
||
const bfd_target bfd_elf32_rx_be_ns_vec;
|
||
const bfd_target bfd_elf32_rx_be_vec;
|
||
|
||
#ifdef DEBUG
|
||
char * rx_get_reloc (long);
|
||
void rx_dump_symtab (bfd *, void *, void *);
|
||
#endif
|
||
|
||
#define RXREL(n,sz,bit,shift,complain,pcrel) \
|
||
HOWTO (R_RX_##n, shift, sz, bit, pcrel, 0, complain_overflow_ ## complain, \
|
||
bfd_elf_generic_reloc, "R_RX_" #n, FALSE, 0, ~0, FALSE)
|
||
|
||
/* Note that the relocations around 0x7f are internal to this file;
|
||
feel free to move them as needed to avoid conflicts with published
|
||
relocation numbers. */
|
||
|
||
static reloc_howto_type rx_elf_howto_table [] =
|
||
{
|
||
RXREL (NONE, 0, 0, 0, dont, FALSE),
|
||
RXREL (DIR32, 2, 32, 0, signed, FALSE),
|
||
RXREL (DIR24S, 2, 24, 0, signed, FALSE),
|
||
RXREL (DIR16, 1, 16, 0, dont, FALSE),
|
||
RXREL (DIR16U, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (DIR16S, 1, 16, 0, signed, FALSE),
|
||
RXREL (DIR8, 0, 8, 0, dont, FALSE),
|
||
RXREL (DIR8U, 0, 8, 0, unsigned, FALSE),
|
||
RXREL (DIR8S, 0, 8, 0, signed, FALSE),
|
||
RXREL (DIR24S_PCREL, 2, 24, 0, signed, TRUE),
|
||
RXREL (DIR16S_PCREL, 1, 16, 0, signed, TRUE),
|
||
RXREL (DIR8S_PCREL, 0, 8, 0, signed, TRUE),
|
||
RXREL (DIR16UL, 1, 16, 2, unsigned, FALSE),
|
||
RXREL (DIR16UW, 1, 16, 1, unsigned, FALSE),
|
||
RXREL (DIR8UL, 0, 8, 2, unsigned, FALSE),
|
||
RXREL (DIR8UW, 0, 8, 1, unsigned, FALSE),
|
||
RXREL (DIR32_REV, 1, 16, 0, dont, FALSE),
|
||
RXREL (DIR16_REV, 1, 16, 0, dont, FALSE),
|
||
RXREL (DIR3U_PCREL, 0, 3, 0, dont, TRUE),
|
||
|
||
EMPTY_HOWTO (0x13),
|
||
EMPTY_HOWTO (0x14),
|
||
EMPTY_HOWTO (0x15),
|
||
EMPTY_HOWTO (0x16),
|
||
EMPTY_HOWTO (0x17),
|
||
EMPTY_HOWTO (0x18),
|
||
EMPTY_HOWTO (0x19),
|
||
EMPTY_HOWTO (0x1a),
|
||
EMPTY_HOWTO (0x1b),
|
||
EMPTY_HOWTO (0x1c),
|
||
EMPTY_HOWTO (0x1d),
|
||
EMPTY_HOWTO (0x1e),
|
||
EMPTY_HOWTO (0x1f),
|
||
|
||
RXREL (RH_3_PCREL, 0, 3, 0, signed, TRUE),
|
||
RXREL (RH_16_OP, 1, 16, 0, signed, FALSE),
|
||
RXREL (RH_24_OP, 2, 24, 0, signed, FALSE),
|
||
RXREL (RH_32_OP, 2, 32, 0, signed, FALSE),
|
||
RXREL (RH_24_UNS, 2, 24, 0, unsigned, FALSE),
|
||
RXREL (RH_8_NEG, 0, 8, 0, signed, FALSE),
|
||
RXREL (RH_16_NEG, 1, 16, 0, signed, FALSE),
|
||
RXREL (RH_24_NEG, 2, 24, 0, signed, FALSE),
|
||
RXREL (RH_32_NEG, 2, 32, 0, signed, FALSE),
|
||
RXREL (RH_DIFF, 2, 32, 0, signed, FALSE),
|
||
RXREL (RH_GPRELB, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (RH_GPRELW, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (RH_GPRELL, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (RH_RELAX, 0, 0, 0, dont, FALSE),
|
||
|
||
EMPTY_HOWTO (0x2e),
|
||
EMPTY_HOWTO (0x2f),
|
||
EMPTY_HOWTO (0x30),
|
||
EMPTY_HOWTO (0x31),
|
||
EMPTY_HOWTO (0x32),
|
||
EMPTY_HOWTO (0x33),
|
||
EMPTY_HOWTO (0x34),
|
||
EMPTY_HOWTO (0x35),
|
||
EMPTY_HOWTO (0x36),
|
||
EMPTY_HOWTO (0x37),
|
||
EMPTY_HOWTO (0x38),
|
||
EMPTY_HOWTO (0x39),
|
||
EMPTY_HOWTO (0x3a),
|
||
EMPTY_HOWTO (0x3b),
|
||
EMPTY_HOWTO (0x3c),
|
||
EMPTY_HOWTO (0x3d),
|
||
EMPTY_HOWTO (0x3e),
|
||
EMPTY_HOWTO (0x3f),
|
||
EMPTY_HOWTO (0x40),
|
||
|
||
RXREL (ABS32, 2, 32, 0, dont, FALSE),
|
||
RXREL (ABS24S, 2, 24, 0, signed, FALSE),
|
||
RXREL (ABS16, 1, 16, 0, dont, FALSE),
|
||
RXREL (ABS16U, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (ABS16S, 1, 16, 0, signed, FALSE),
|
||
RXREL (ABS8, 0, 8, 0, dont, FALSE),
|
||
RXREL (ABS8U, 0, 8, 0, unsigned, FALSE),
|
||
RXREL (ABS8S, 0, 8, 0, signed, FALSE),
|
||
RXREL (ABS24S_PCREL, 2, 24, 0, signed, TRUE),
|
||
RXREL (ABS16S_PCREL, 1, 16, 0, signed, TRUE),
|
||
RXREL (ABS8S_PCREL, 0, 8, 0, signed, TRUE),
|
||
RXREL (ABS16UL, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (ABS16UW, 1, 16, 0, unsigned, FALSE),
|
||
RXREL (ABS8UL, 0, 8, 0, unsigned, FALSE),
|
||
RXREL (ABS8UW, 0, 8, 0, unsigned, FALSE),
|
||
RXREL (ABS32_REV, 2, 32, 0, dont, FALSE),
|
||
RXREL (ABS16_REV, 1, 16, 0, dont, FALSE),
|
||
|
||
#define STACK_REL_P(x) ((x) <= R_RX_ABS16_REV && (x) >= R_RX_ABS32)
|
||
|
||
EMPTY_HOWTO (0x52),
|
||
EMPTY_HOWTO (0x53),
|
||
EMPTY_HOWTO (0x54),
|
||
EMPTY_HOWTO (0x55),
|
||
EMPTY_HOWTO (0x56),
|
||
EMPTY_HOWTO (0x57),
|
||
EMPTY_HOWTO (0x58),
|
||
EMPTY_HOWTO (0x59),
|
||
EMPTY_HOWTO (0x5a),
|
||
EMPTY_HOWTO (0x5b),
|
||
EMPTY_HOWTO (0x5c),
|
||
EMPTY_HOWTO (0x5d),
|
||
EMPTY_HOWTO (0x5e),
|
||
EMPTY_HOWTO (0x5f),
|
||
EMPTY_HOWTO (0x60),
|
||
EMPTY_HOWTO (0x61),
|
||
EMPTY_HOWTO (0x62),
|
||
EMPTY_HOWTO (0x63),
|
||
EMPTY_HOWTO (0x64),
|
||
EMPTY_HOWTO (0x65),
|
||
EMPTY_HOWTO (0x66),
|
||
EMPTY_HOWTO (0x67),
|
||
EMPTY_HOWTO (0x68),
|
||
EMPTY_HOWTO (0x69),
|
||
EMPTY_HOWTO (0x6a),
|
||
EMPTY_HOWTO (0x6b),
|
||
EMPTY_HOWTO (0x6c),
|
||
EMPTY_HOWTO (0x6d),
|
||
EMPTY_HOWTO (0x6e),
|
||
EMPTY_HOWTO (0x6f),
|
||
EMPTY_HOWTO (0x70),
|
||
EMPTY_HOWTO (0x71),
|
||
EMPTY_HOWTO (0x72),
|
||
EMPTY_HOWTO (0x73),
|
||
EMPTY_HOWTO (0x74),
|
||
EMPTY_HOWTO (0x75),
|
||
EMPTY_HOWTO (0x76),
|
||
EMPTY_HOWTO (0x77),
|
||
|
||
/* These are internal. */
|
||
/* A 5-bit unsigned displacement to a B/W/L address, at bit position 8/12. */
|
||
/* ---- ---- 4--- 3210. */
|
||
#define R_RX_RH_ABS5p8B 0x78
|
||
RXREL (RH_ABS5p8B, 0, 0, 0, dont, FALSE),
|
||
#define R_RX_RH_ABS5p8W 0x79
|
||
RXREL (RH_ABS5p8W, 0, 0, 0, dont, FALSE),
|
||
#define R_RX_RH_ABS5p8L 0x7a
|
||
RXREL (RH_ABS5p8L, 0, 0, 0, dont, FALSE),
|
||
/* A 5-bit unsigned displacement to a B/W/L address, at bit position 5/12. */
|
||
/* ---- -432 1--- 0---. */
|
||
#define R_RX_RH_ABS5p5B 0x7b
|
||
RXREL (RH_ABS5p5B, 0, 0, 0, dont, FALSE),
|
||
#define R_RX_RH_ABS5p5W 0x7c
|
||
RXREL (RH_ABS5p5W, 0, 0, 0, dont, FALSE),
|
||
#define R_RX_RH_ABS5p5L 0x7d
|
||
RXREL (RH_ABS5p5L, 0, 0, 0, dont, FALSE),
|
||
/* A 4-bit unsigned immediate at bit position 8. */
|
||
#define R_RX_RH_UIMM4p8 0x7e
|
||
RXREL (RH_UIMM4p8, 0, 0, 0, dont, FALSE),
|
||
/* A 4-bit negative unsigned immediate at bit position 8. */
|
||
#define R_RX_RH_UNEG4p8 0x7f
|
||
RXREL (RH_UNEG4p8, 0, 0, 0, dont, FALSE),
|
||
/* End of internal relocs. */
|
||
|
||
RXREL (SYM, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPneg, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPadd, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPsub, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPmul, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPdiv, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPshla, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPshra, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPsctsize, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPscttop, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPand, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPor, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPxor, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPnot, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPmod, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPromtop, 2, 32, 0, dont, FALSE),
|
||
RXREL (OPramtop, 2, 32, 0, dont, FALSE)
|
||
};
|
||
|
||
/* Map BFD reloc types to RX ELF reloc types. */
|
||
|
||
struct rx_reloc_map
|
||
{
|
||
bfd_reloc_code_real_type bfd_reloc_val;
|
||
unsigned int rx_reloc_val;
|
||
};
|
||
|
||
static const struct rx_reloc_map rx_reloc_map [] =
|
||
{
|
||
{ BFD_RELOC_NONE, R_RX_NONE },
|
||
{ BFD_RELOC_8, R_RX_DIR8S },
|
||
{ BFD_RELOC_16, R_RX_DIR16S },
|
||
{ BFD_RELOC_24, R_RX_DIR24S },
|
||
{ BFD_RELOC_32, R_RX_DIR32 },
|
||
{ BFD_RELOC_RX_16_OP, R_RX_DIR16 },
|
||
{ BFD_RELOC_RX_DIR3U_PCREL, R_RX_DIR3U_PCREL },
|
||
{ BFD_RELOC_8_PCREL, R_RX_DIR8S_PCREL },
|
||
{ BFD_RELOC_16_PCREL, R_RX_DIR16S_PCREL },
|
||
{ BFD_RELOC_24_PCREL, R_RX_DIR24S_PCREL },
|
||
{ BFD_RELOC_RX_8U, R_RX_DIR8U },
|
||
{ BFD_RELOC_RX_16U, R_RX_DIR16U },
|
||
{ BFD_RELOC_RX_24U, R_RX_RH_24_UNS },
|
||
{ BFD_RELOC_RX_NEG8, R_RX_RH_8_NEG },
|
||
{ BFD_RELOC_RX_NEG16, R_RX_RH_16_NEG },
|
||
{ BFD_RELOC_RX_NEG24, R_RX_RH_24_NEG },
|
||
{ BFD_RELOC_RX_NEG32, R_RX_RH_32_NEG },
|
||
{ BFD_RELOC_RX_DIFF, R_RX_RH_DIFF },
|
||
{ BFD_RELOC_RX_GPRELB, R_RX_RH_GPRELB },
|
||
{ BFD_RELOC_RX_GPRELW, R_RX_RH_GPRELW },
|
||
{ BFD_RELOC_RX_GPRELL, R_RX_RH_GPRELL },
|
||
{ BFD_RELOC_RX_RELAX, R_RX_RH_RELAX },
|
||
{ BFD_RELOC_RX_SYM, R_RX_SYM },
|
||
{ BFD_RELOC_RX_OP_SUBTRACT, R_RX_OPsub },
|
||
{ BFD_RELOC_RX_OP_NEG, R_RX_OPneg },
|
||
{ BFD_RELOC_RX_ABS8, R_RX_ABS8 },
|
||
{ BFD_RELOC_RX_ABS16, R_RX_ABS16 },
|
||
{ BFD_RELOC_RX_ABS16_REV, R_RX_ABS16_REV },
|
||
{ BFD_RELOC_RX_ABS32, R_RX_ABS32 },
|
||
{ BFD_RELOC_RX_ABS32_REV, R_RX_ABS32_REV },
|
||
{ BFD_RELOC_RX_ABS16UL, R_RX_ABS16UL },
|
||
{ BFD_RELOC_RX_ABS16UW, R_RX_ABS16UW },
|
||
{ BFD_RELOC_RX_ABS16U, R_RX_ABS16U }
|
||
};
|
||
|
||
#define BIGE(abfd) ((abfd)->xvec->byteorder == BFD_ENDIAN_BIG)
|
||
|
||
static reloc_howto_type *
|
||
rx_reloc_type_lookup (bfd * abfd ATTRIBUTE_UNUSED,
|
||
bfd_reloc_code_real_type code)
|
||
{
|
||
unsigned int i;
|
||
|
||
if (code == BFD_RELOC_RX_32_OP)
|
||
return rx_elf_howto_table + R_RX_DIR32;
|
||
|
||
for (i = ARRAY_SIZE (rx_reloc_map); --i;)
|
||
if (rx_reloc_map [i].bfd_reloc_val == code)
|
||
return rx_elf_howto_table + rx_reloc_map[i].rx_reloc_val;
|
||
|
||
return NULL;
|
||
}
|
||
|
||
static reloc_howto_type *
|
||
rx_reloc_name_lookup (bfd * abfd ATTRIBUTE_UNUSED, const char * r_name)
|
||
{
|
||
unsigned int i;
|
||
|
||
for (i = 0; i < ARRAY_SIZE (rx_elf_howto_table); i++)
|
||
if (rx_elf_howto_table[i].name != NULL
|
||
&& strcasecmp (rx_elf_howto_table[i].name, r_name) == 0)
|
||
return rx_elf_howto_table + i;
|
||
|
||
return NULL;
|
||
}
|
||
|
||
/* Set the howto pointer for an RX ELF reloc. */
|
||
|
||
static void
|
||
rx_info_to_howto_rela (bfd * abfd ATTRIBUTE_UNUSED,
|
||
arelent * cache_ptr,
|
||
Elf_Internal_Rela * dst)
|
||
{
|
||
unsigned int r_type;
|
||
|
||
r_type = ELF32_R_TYPE (dst->r_info);
|
||
BFD_ASSERT (r_type < (unsigned int) R_RX_max);
|
||
cache_ptr->howto = rx_elf_howto_table + r_type;
|
||
}
|
||
|
||
static bfd_vma
|
||
get_symbol_value (const char * name,
|
||
bfd_reloc_status_type * status,
|
||
struct bfd_link_info * info,
|
||
bfd * input_bfd,
|
||
asection * input_section,
|
||
int offset)
|
||
{
|
||
bfd_vma value = 0;
|
||
struct bfd_link_hash_entry * h;
|
||
|
||
h = bfd_link_hash_lookup (info->hash, name, FALSE, FALSE, TRUE);
|
||
|
||
if (h == NULL
|
||
|| (h->type != bfd_link_hash_defined
|
||
&& h->type != bfd_link_hash_defweak))
|
||
* status = info->callbacks->undefined_symbol
|
||
(info, name, input_bfd, input_section, offset, TRUE);
|
||
else
|
||
value = (h->u.def.value
|
||
+ h->u.def.section->output_section->vma
|
||
+ h->u.def.section->output_offset);
|
||
|
||
return value;
|
||
}
|
||
|
||
static bfd_vma
|
||
get_gp (bfd_reloc_status_type * status,
|
||
struct bfd_link_info * info,
|
||
bfd * abfd,
|
||
asection * sec,
|
||
int offset)
|
||
{
|
||
static bfd_boolean cached = FALSE;
|
||
static bfd_vma cached_value = 0;
|
||
|
||
if (!cached)
|
||
{
|
||
cached_value = get_symbol_value ("__gp", status, info, abfd, sec, offset);
|
||
cached = TRUE;
|
||
}
|
||
return cached_value;
|
||
}
|
||
|
||
static bfd_vma
|
||
get_romstart (bfd_reloc_status_type * status,
|
||
struct bfd_link_info * info,
|
||
bfd * abfd,
|
||
asection * sec,
|
||
int offset)
|
||
{
|
||
static bfd_boolean cached = FALSE;
|
||
static bfd_vma cached_value = 0;
|
||
|
||
if (!cached)
|
||
{
|
||
cached_value = get_symbol_value ("_start", status, info, abfd, sec, offset);
|
||
cached = TRUE;
|
||
}
|
||
return cached_value;
|
||
}
|
||
|
||
static bfd_vma
|
||
get_ramstart (bfd_reloc_status_type * status,
|
||
struct bfd_link_info * info,
|
||
bfd * abfd,
|
||
asection * sec,
|
||
int offset)
|
||
{
|
||
static bfd_boolean cached = FALSE;
|
||
static bfd_vma cached_value = 0;
|
||
|
||
if (!cached)
|
||
{
|
||
cached_value = get_symbol_value ("__datastart", status, info, abfd, sec, offset);
|
||
cached = TRUE;
|
||
}
|
||
return cached_value;
|
||
}
|
||
|
||
#define NUM_STACK_ENTRIES 16
|
||
static int32_t rx_stack [ NUM_STACK_ENTRIES ];
|
||
static unsigned int rx_stack_top;
|
||
|
||
#define RX_STACK_PUSH(val) \
|
||
do \
|
||
{ \
|
||
if (rx_stack_top < NUM_STACK_ENTRIES) \
|
||
rx_stack [rx_stack_top ++] = (val); \
|
||
else \
|
||
r = bfd_reloc_dangerous; \
|
||
} \
|
||
while (0)
|
||
|
||
#define RX_STACK_POP(dest) \
|
||
do \
|
||
{ \
|
||
if (rx_stack_top > 0) \
|
||
(dest) = rx_stack [-- rx_stack_top]; \
|
||
else \
|
||
(dest) = 0, r = bfd_reloc_dangerous; \
|
||
} \
|
||
while (0)
|
||
|
||
/* Relocate an RX ELF section.
|
||
There is some attempt to make this function usable for many architectures,
|
||
both USE_REL and USE_RELA ['twould be nice if such a critter existed],
|
||
if only to serve as a learning tool.
|
||
|
||
The RELOCATE_SECTION function is called by the new ELF backend linker
|
||
to handle the relocations for a section.
|
||
|
||
The relocs are always passed as Rela structures; if the section
|
||
actually uses Rel structures, the r_addend field will always be
|
||
zero.
|
||
|
||
This function is responsible for adjusting the section contents as
|
||
necessary, and (if using Rela relocs and generating a relocatable
|
||
output file) adjusting the reloc addend as necessary.
|
||
|
||
This function does not have to worry about setting the reloc
|
||
address or the reloc symbol index.
|
||
|
||
LOCAL_SYMS is a pointer to the swapped in local symbols.
|
||
|
||
LOCAL_SECTIONS is an array giving the section in the input file
|
||
corresponding to the st_shndx field of each local symbol.
|
||
|
||
The global hash table entry for the global symbols can be found
|
||
via elf_sym_hashes (input_bfd).
|
||
|
||
When generating relocatable output, this function must handle
|
||
STB_LOCAL/STT_SECTION symbols specially. The output symbol is
|
||
going to be the section symbol corresponding to the output
|
||
section, which means that the addend must be adjusted
|
||
accordingly. */
|
||
|
||
static bfd_boolean
|
||
rx_elf_relocate_section
|
||
(bfd * output_bfd,
|
||
struct bfd_link_info * info,
|
||
bfd * input_bfd,
|
||
asection * input_section,
|
||
bfd_byte * contents,
|
||
Elf_Internal_Rela * relocs,
|
||
Elf_Internal_Sym * local_syms,
|
||
asection ** local_sections)
|
||
{
|
||
Elf_Internal_Shdr * symtab_hdr;
|
||
struct elf_link_hash_entry ** sym_hashes;
|
||
Elf_Internal_Rela * rel;
|
||
Elf_Internal_Rela * relend;
|
||
bfd_boolean pid_mode;
|
||
bfd_boolean saw_subtract = FALSE;
|
||
|
||
if (elf_elfheader (output_bfd)->e_flags & E_FLAG_RX_PID)
|
||
pid_mode = TRUE;
|
||
else
|
||
pid_mode = FALSE;
|
||
|
||
symtab_hdr = & elf_tdata (input_bfd)->symtab_hdr;
|
||
sym_hashes = elf_sym_hashes (input_bfd);
|
||
relend = relocs + input_section->reloc_count;
|
||
for (rel = relocs; rel < relend; rel ++)
|
||
{
|
||
reloc_howto_type * howto;
|
||
unsigned long r_symndx;
|
||
Elf_Internal_Sym * sym;
|
||
asection * sec;
|
||
struct elf_link_hash_entry * h;
|
||
bfd_vma relocation;
|
||
bfd_reloc_status_type r;
|
||
const char * name = NULL;
|
||
bfd_boolean unresolved_reloc = TRUE;
|
||
int r_type;
|
||
|
||
r_type = ELF32_R_TYPE (rel->r_info);
|
||
r_symndx = ELF32_R_SYM (rel->r_info);
|
||
|
||
howto = rx_elf_howto_table + ELF32_R_TYPE (rel->r_info);
|
||
h = NULL;
|
||
sym = NULL;
|
||
sec = NULL;
|
||
relocation = 0;
|
||
|
||
if (rx_stack_top == 0)
|
||
saw_subtract = FALSE;
|
||
|
||
if (r_symndx < symtab_hdr->sh_info)
|
||
{
|
||
sym = local_syms + r_symndx;
|
||
sec = local_sections [r_symndx];
|
||
relocation = _bfd_elf_rela_local_sym (output_bfd, sym, & sec, rel);
|
||
|
||
name = bfd_elf_string_from_elf_section
|
||
(input_bfd, symtab_hdr->sh_link, sym->st_name);
|
||
name = (sym->st_name == 0) ? bfd_section_name (input_bfd, sec) : name;
|
||
}
|
||
else
|
||
{
|
||
bfd_boolean warned;
|
||
|
||
RELOC_FOR_GLOBAL_SYMBOL (info, input_bfd, input_section, rel,
|
||
r_symndx, symtab_hdr, sym_hashes, h,
|
||
sec, relocation, unresolved_reloc,
|
||
warned);
|
||
|
||
name = h->root.root.string;
|
||
}
|
||
|
||
if (sec != NULL && elf_discarded_section (sec))
|
||
RELOC_AGAINST_DISCARDED_SECTION (info, input_bfd, input_section,
|
||
rel, relend, howto, contents);
|
||
|
||
if (info->relocatable)
|
||
{
|
||
/* This is a relocatable link. We don't have to change
|
||
anything, unless the reloc is against a section symbol,
|
||
in which case we have to adjust according to where the
|
||
section symbol winds up in the output section. */
|
||
if (sym != NULL && ELF_ST_TYPE (sym->st_info) == STT_SECTION)
|
||
rel->r_addend += sec->output_offset;
|
||
continue;
|
||
}
|
||
|
||
if (h != NULL && h->root.type == bfd_link_hash_undefweak)
|
||
/* If the symbol is undefined and weak
|
||
then the relocation resolves to zero. */
|
||
relocation = 0;
|
||
else
|
||
{
|
||
if (howto->pc_relative)
|
||
{
|
||
relocation -= (input_section->output_section->vma
|
||
+ input_section->output_offset
|
||
+ rel->r_offset);
|
||
if (r_type != R_RX_RH_3_PCREL
|
||
&& r_type != R_RX_DIR3U_PCREL)
|
||
relocation ++;
|
||
}
|
||
|
||
relocation += rel->r_addend;
|
||
}
|
||
|
||
r = bfd_reloc_ok;
|
||
|
||
#define RANGE(a,b) if (a > (long) relocation || (long) relocation > b) r = bfd_reloc_overflow
|
||
#define ALIGN(m) if (relocation & m) r = bfd_reloc_other;
|
||
#define OP(i) (contents[rel->r_offset + (i)])
|
||
#define WARN_REDHAT(type) \
|
||
_bfd_error_handler (_("%B:%A: Warning: deprecated Red Hat reloc " type " detected against: %s."), \
|
||
input_bfd, input_section, name)
|
||
|
||
/* Check for unsafe relocs in PID mode. These are any relocs where
|
||
an absolute address is being computed. There are special cases
|
||
for relocs against symbols that are known to be referenced in
|
||
crt0.o before the PID base address register has been initialised. */
|
||
#define UNSAFE_FOR_PID \
|
||
do \
|
||
{ \
|
||
if (pid_mode \
|
||
&& sec != NULL \
|
||
&& sec->flags & SEC_READONLY \
|
||
&& !(input_section->flags & SEC_DEBUGGING) \
|
||
&& strcmp (name, "__pid_base") != 0 \
|
||
&& strcmp (name, "__gp") != 0 \
|
||
&& strcmp (name, "__romdatastart") != 0 \
|
||
&& !saw_subtract) \
|
||
_bfd_error_handler (_("%B(%A): unsafe PID relocation %s at 0x%08lx (against %s in %s)"), \
|
||
input_bfd, input_section, howto->name, \
|
||
input_section->output_section->vma + input_section->output_offset + rel->r_offset, \
|
||
name, sec->name); \
|
||
} \
|
||
while (0)
|
||
|
||
/* Opcode relocs are always big endian. Data relocs are bi-endian. */
|
||
switch (r_type)
|
||
{
|
||
case R_RX_NONE:
|
||
break;
|
||
|
||
case R_RX_RH_RELAX:
|
||
break;
|
||
|
||
case R_RX_RH_3_PCREL:
|
||
WARN_REDHAT ("RX_RH_3_PCREL");
|
||
RANGE (3, 10);
|
||
OP (0) &= 0xf8;
|
||
OP (0) |= relocation & 0x07;
|
||
break;
|
||
|
||
case R_RX_RH_8_NEG:
|
||
WARN_REDHAT ("RX_RH_8_NEG");
|
||
relocation = - relocation;
|
||
case R_RX_DIR8S_PCREL:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-128, 127);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_DIR8S:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-128, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_DIR8U:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (0, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_RH_16_NEG:
|
||
WARN_REDHAT ("RX_RH_16_NEG");
|
||
relocation = - relocation;
|
||
case R_RX_DIR16S_PCREL:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-32768, 32767);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_16_OP:
|
||
WARN_REDHAT ("RX_RH_16_OP");
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-32768, 32767);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR16S:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-32768, 65535);
|
||
if (BIGE (output_bfd) && !(input_section->flags & SEC_CODE))
|
||
{
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
}
|
||
else
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
}
|
||
break;
|
||
|
||
case R_RX_DIR16U:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (0, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR16:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-32768, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR16_REV:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-32768, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#else
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR3U_PCREL:
|
||
RANGE (3, 10);
|
||
OP (0) &= 0xf8;
|
||
OP (0) |= relocation & 0x07;
|
||
break;
|
||
|
||
case R_RX_RH_24_NEG:
|
||
UNSAFE_FOR_PID;
|
||
WARN_REDHAT ("RX_RH_24_NEG");
|
||
relocation = - relocation;
|
||
case R_RX_DIR24S_PCREL:
|
||
RANGE (-0x800000, 0x7fffff);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (2) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (0) = relocation >> 16;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_24_OP:
|
||
UNSAFE_FOR_PID;
|
||
WARN_REDHAT ("RX_RH_24_OP");
|
||
RANGE (-0x800000, 0x7fffff);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (2) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (0) = relocation >> 16;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR24S:
|
||
UNSAFE_FOR_PID;
|
||
RANGE (-0x800000, 0x7fffff);
|
||
if (BIGE (output_bfd) && !(input_section->flags & SEC_CODE))
|
||
{
|
||
OP (2) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (0) = relocation >> 16;
|
||
}
|
||
else
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
}
|
||
break;
|
||
|
||
case R_RX_RH_24_UNS:
|
||
UNSAFE_FOR_PID;
|
||
WARN_REDHAT ("RX_RH_24_UNS");
|
||
RANGE (0, 0xffffff);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (2) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (0) = relocation >> 16;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_32_NEG:
|
||
UNSAFE_FOR_PID;
|
||
WARN_REDHAT ("RX_RH_32_NEG");
|
||
relocation = - relocation;
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_32_OP:
|
||
UNSAFE_FOR_PID;
|
||
WARN_REDHAT ("RX_RH_32_OP");
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_DIR32:
|
||
if (BIGE (output_bfd) && !(input_section->flags & SEC_CODE))
|
||
{
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
}
|
||
else
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
}
|
||
break;
|
||
|
||
case R_RX_DIR32_REV:
|
||
if (BIGE (output_bfd))
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
}
|
||
else
|
||
{
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
}
|
||
break;
|
||
|
||
case R_RX_RH_DIFF:
|
||
{
|
||
bfd_vma val;
|
||
WARN_REDHAT ("RX_RH_DIFF");
|
||
val = bfd_get_32 (output_bfd, & OP (0));
|
||
val -= relocation;
|
||
bfd_put_32 (output_bfd, val, & OP (0));
|
||
}
|
||
break;
|
||
|
||
case R_RX_RH_GPRELB:
|
||
WARN_REDHAT ("RX_RH_GPRELB");
|
||
relocation -= get_gp (&r, info, input_bfd, input_section, rel->r_offset);
|
||
RANGE (0, 65535);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_GPRELW:
|
||
WARN_REDHAT ("RX_RH_GPRELW");
|
||
relocation -= get_gp (&r, info, input_bfd, input_section, rel->r_offset);
|
||
ALIGN (1);
|
||
relocation >>= 1;
|
||
RANGE (0, 65535);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_RH_GPRELL:
|
||
WARN_REDHAT ("RX_RH_GPRELL");
|
||
relocation -= get_gp (&r, info, input_bfd, input_section, rel->r_offset);
|
||
ALIGN (3);
|
||
relocation >>= 2;
|
||
RANGE (0, 65535);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
/* Internal relocations just for relaxation: */
|
||
case R_RX_RH_ABS5p5B:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 31);
|
||
OP (0) &= 0xf8;
|
||
OP (0) |= relocation >> 2;
|
||
OP (1) &= 0x77;
|
||
OP (1) |= (relocation << 6) & 0x80;
|
||
OP (1) |= (relocation << 3) & 0x08;
|
||
break;
|
||
|
||
case R_RX_RH_ABS5p5W:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 62);
|
||
ALIGN (1);
|
||
relocation >>= 1;
|
||
OP (0) &= 0xf8;
|
||
OP (0) |= relocation >> 2;
|
||
OP (1) &= 0x77;
|
||
OP (1) |= (relocation << 6) & 0x80;
|
||
OP (1) |= (relocation << 3) & 0x08;
|
||
break;
|
||
|
||
case R_RX_RH_ABS5p5L:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 124);
|
||
ALIGN (3);
|
||
relocation >>= 2;
|
||
OP (0) &= 0xf8;
|
||
OP (0) |= relocation >> 2;
|
||
OP (1) &= 0x77;
|
||
OP (1) |= (relocation << 6) & 0x80;
|
||
OP (1) |= (relocation << 3) & 0x08;
|
||
break;
|
||
|
||
case R_RX_RH_ABS5p8B:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 31);
|
||
OP (0) &= 0x70;
|
||
OP (0) |= (relocation << 3) & 0x80;
|
||
OP (0) |= relocation & 0x0f;
|
||
break;
|
||
|
||
case R_RX_RH_ABS5p8W:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 62);
|
||
ALIGN (1);
|
||
relocation >>= 1;
|
||
OP (0) &= 0x70;
|
||
OP (0) |= (relocation << 3) & 0x80;
|
||
OP (0) |= relocation & 0x0f;
|
||
break;
|
||
|
||
case R_RX_RH_ABS5p8L:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 124);
|
||
ALIGN (3);
|
||
relocation >>= 2;
|
||
OP (0) &= 0x70;
|
||
OP (0) |= (relocation << 3) & 0x80;
|
||
OP (0) |= relocation & 0x0f;
|
||
break;
|
||
|
||
case R_RX_RH_UIMM4p8:
|
||
RANGE (0, 15);
|
||
OP (0) &= 0x0f;
|
||
OP (0) |= relocation << 4;
|
||
break;
|
||
|
||
case R_RX_RH_UNEG4p8:
|
||
RANGE (-15, 0);
|
||
OP (0) &= 0x0f;
|
||
OP (0) |= (-relocation) << 4;
|
||
break;
|
||
|
||
/* Complex reloc handling: */
|
||
|
||
case R_RX_ABS32:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS32_REV:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
OP (3) = relocation >> 24;
|
||
#else
|
||
OP (3) = relocation;
|
||
OP (2) = relocation >> 8;
|
||
OP (1) = relocation >> 16;
|
||
OP (0) = relocation >> 24;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS24S_PCREL:
|
||
case R_RX_ABS24S:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-0x800000, 0x7fffff);
|
||
if (BIGE (output_bfd) && !(input_section->flags & SEC_CODE))
|
||
{
|
||
OP (2) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (0) = relocation >> 16;
|
||
}
|
||
else
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
OP (2) = relocation >> 16;
|
||
}
|
||
break;
|
||
|
||
case R_RX_ABS16:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-32768, 65535);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS16_REV:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-32768, 65535);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#else
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS16S_PCREL:
|
||
case R_RX_ABS16S:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-32768, 32767);
|
||
if (BIGE (output_bfd) && !(input_section->flags & SEC_CODE))
|
||
{
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
}
|
||
else
|
||
{
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
}
|
||
break;
|
||
|
||
case R_RX_ABS16U:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS16UL:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
relocation >>= 2;
|
||
RANGE (0, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS16UW:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
relocation >>= 1;
|
||
RANGE (0, 65536);
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
OP (1) = relocation;
|
||
OP (0) = relocation >> 8;
|
||
#else
|
||
OP (0) = relocation;
|
||
OP (1) = relocation >> 8;
|
||
#endif
|
||
break;
|
||
|
||
case R_RX_ABS8:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-128, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_ABS8U:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
RANGE (0, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_ABS8UL:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
relocation >>= 2;
|
||
RANGE (0, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_ABS8UW:
|
||
UNSAFE_FOR_PID;
|
||
RX_STACK_POP (relocation);
|
||
relocation >>= 1;
|
||
RANGE (0, 255);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_ABS8S:
|
||
UNSAFE_FOR_PID;
|
||
case R_RX_ABS8S_PCREL:
|
||
RX_STACK_POP (relocation);
|
||
RANGE (-128, 127);
|
||
OP (0) = relocation;
|
||
break;
|
||
|
||
case R_RX_SYM:
|
||
if (r_symndx < symtab_hdr->sh_info)
|
||
RX_STACK_PUSH (sec->output_section->vma
|
||
+ sec->output_offset
|
||
+ sym->st_value
|
||
+ rel->r_addend);
|
||
else
|
||
{
|
||
if (h != NULL
|
||
&& (h->root.type == bfd_link_hash_defined
|
||
|| h->root.type == bfd_link_hash_defweak))
|
||
RX_STACK_PUSH (h->root.u.def.value
|
||
+ sec->output_section->vma
|
||
+ sec->output_offset
|
||
+ rel->r_addend);
|
||
else
|
||
_bfd_error_handler (_("Warning: RX_SYM reloc with an unknown symbol"));
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPneg:
|
||
{
|
||
int32_t tmp;
|
||
|
||
RX_STACK_POP (tmp);
|
||
tmp = - tmp;
|
||
RX_STACK_PUSH (tmp);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPadd:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 += tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPsub:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
saw_subtract = TRUE;
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp2 -= tmp1;
|
||
RX_STACK_PUSH (tmp2);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPmul:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 *= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPdiv:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
saw_subtract = TRUE;
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 /= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPshla:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 <<= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPshra:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 >>= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPsctsize:
|
||
RX_STACK_PUSH (input_section->size);
|
||
break;
|
||
|
||
case R_RX_OPscttop:
|
||
RX_STACK_PUSH (input_section->output_section->vma);
|
||
break;
|
||
|
||
case R_RX_OPand:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 &= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPor:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 |= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPxor:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 ^= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPnot:
|
||
{
|
||
int32_t tmp;
|
||
|
||
RX_STACK_POP (tmp);
|
||
tmp = ~ tmp;
|
||
RX_STACK_PUSH (tmp);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPmod:
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 %= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
}
|
||
break;
|
||
|
||
case R_RX_OPromtop:
|
||
RX_STACK_PUSH (get_romstart (&r, info, input_bfd, input_section, rel->r_offset));
|
||
break;
|
||
|
||
case R_RX_OPramtop:
|
||
RX_STACK_PUSH (get_ramstart (&r, info, input_bfd, input_section, rel->r_offset));
|
||
break;
|
||
|
||
default:
|
||
r = bfd_reloc_notsupported;
|
||
break;
|
||
}
|
||
|
||
if (r != bfd_reloc_ok)
|
||
{
|
||
const char * msg = NULL;
|
||
|
||
switch (r)
|
||
{
|
||
case bfd_reloc_overflow:
|
||
/* Catch the case of a missing function declaration
|
||
and emit a more helpful error message. */
|
||
if (r_type == R_RX_DIR24S_PCREL)
|
||
msg = _("%B(%A): error: call to undefined function '%s'");
|
||
else
|
||
r = info->callbacks->reloc_overflow
|
||
(info, (h ? &h->root : NULL), name, howto->name, (bfd_vma) 0,
|
||
input_bfd, input_section, rel->r_offset);
|
||
break;
|
||
|
||
case bfd_reloc_undefined:
|
||
r = info->callbacks->undefined_symbol
|
||
(info, name, input_bfd, input_section, rel->r_offset,
|
||
TRUE);
|
||
break;
|
||
|
||
case bfd_reloc_other:
|
||
msg = _("%B(%A): warning: unaligned access to symbol '%s' in the small data area");
|
||
break;
|
||
|
||
case bfd_reloc_outofrange:
|
||
msg = _("%B(%A): internal error: out of range error");
|
||
break;
|
||
|
||
case bfd_reloc_notsupported:
|
||
msg = _("%B(%A): internal error: unsupported relocation error");
|
||
break;
|
||
|
||
case bfd_reloc_dangerous:
|
||
msg = _("%B(%A): internal error: dangerous relocation");
|
||
break;
|
||
|
||
default:
|
||
msg = _("%B(%A): internal error: unknown error");
|
||
break;
|
||
}
|
||
|
||
if (msg)
|
||
_bfd_error_handler (msg, input_bfd, input_section, name);
|
||
|
||
if (! r)
|
||
return FALSE;
|
||
}
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
/* Relaxation Support. */
|
||
|
||
/* Progression of relocations from largest operand size to smallest
|
||
operand size. */
|
||
|
||
static int
|
||
next_smaller_reloc (int r)
|
||
{
|
||
switch (r)
|
||
{
|
||
case R_RX_DIR32: return R_RX_DIR24S;
|
||
case R_RX_DIR24S: return R_RX_DIR16S;
|
||
case R_RX_DIR16S: return R_RX_DIR8S;
|
||
case R_RX_DIR8S: return R_RX_NONE;
|
||
|
||
case R_RX_DIR16: return R_RX_DIR8;
|
||
case R_RX_DIR8: return R_RX_NONE;
|
||
|
||
case R_RX_DIR16U: return R_RX_DIR8U;
|
||
case R_RX_DIR8U: return R_RX_NONE;
|
||
|
||
case R_RX_DIR24S_PCREL: return R_RX_DIR16S_PCREL;
|
||
case R_RX_DIR16S_PCREL: return R_RX_DIR8S_PCREL;
|
||
case R_RX_DIR8S_PCREL: return R_RX_DIR3U_PCREL;
|
||
|
||
case R_RX_DIR16UL: return R_RX_DIR8UL;
|
||
case R_RX_DIR8UL: return R_RX_NONE;
|
||
case R_RX_DIR16UW: return R_RX_DIR8UW;
|
||
case R_RX_DIR8UW: return R_RX_NONE;
|
||
|
||
case R_RX_RH_32_OP: return R_RX_RH_24_OP;
|
||
case R_RX_RH_24_OP: return R_RX_RH_16_OP;
|
||
case R_RX_RH_16_OP: return R_RX_DIR8;
|
||
|
||
case R_RX_ABS32: return R_RX_ABS24S;
|
||
case R_RX_ABS24S: return R_RX_ABS16S;
|
||
case R_RX_ABS16: return R_RX_ABS8;
|
||
case R_RX_ABS16U: return R_RX_ABS8U;
|
||
case R_RX_ABS16S: return R_RX_ABS8S;
|
||
case R_RX_ABS8: return R_RX_NONE;
|
||
case R_RX_ABS8U: return R_RX_NONE;
|
||
case R_RX_ABS8S: return R_RX_NONE;
|
||
case R_RX_ABS24S_PCREL: return R_RX_ABS16S_PCREL;
|
||
case R_RX_ABS16S_PCREL: return R_RX_ABS8S_PCREL;
|
||
case R_RX_ABS8S_PCREL: return R_RX_NONE;
|
||
case R_RX_ABS16UL: return R_RX_ABS8UL;
|
||
case R_RX_ABS16UW: return R_RX_ABS8UW;
|
||
case R_RX_ABS8UL: return R_RX_NONE;
|
||
case R_RX_ABS8UW: return R_RX_NONE;
|
||
}
|
||
return r;
|
||
};
|
||
|
||
/* Delete some bytes from a section while relaxing. */
|
||
|
||
static bfd_boolean
|
||
elf32_rx_relax_delete_bytes (bfd *abfd, asection *sec, bfd_vma addr, int count,
|
||
Elf_Internal_Rela *alignment_rel, int force_snip)
|
||
{
|
||
Elf_Internal_Shdr * symtab_hdr;
|
||
unsigned int sec_shndx;
|
||
bfd_byte * contents;
|
||
Elf_Internal_Rela * irel;
|
||
Elf_Internal_Rela * irelend;
|
||
Elf_Internal_Sym * isym;
|
||
Elf_Internal_Sym * isymend;
|
||
bfd_vma toaddr;
|
||
unsigned int symcount;
|
||
struct elf_link_hash_entry ** sym_hashes;
|
||
struct elf_link_hash_entry ** end_hashes;
|
||
|
||
if (!alignment_rel)
|
||
force_snip = 1;
|
||
|
||
sec_shndx = _bfd_elf_section_from_bfd_section (abfd, sec);
|
||
|
||
contents = elf_section_data (sec)->this_hdr.contents;
|
||
|
||
/* The deletion must stop at the next alignment boundary, if
|
||
ALIGNMENT_REL is non-NULL. */
|
||
toaddr = sec->size;
|
||
if (alignment_rel)
|
||
toaddr = alignment_rel->r_offset;
|
||
|
||
irel = elf_section_data (sec)->relocs;
|
||
irelend = irel + sec->reloc_count;
|
||
|
||
/* Actually delete the bytes. */
|
||
memmove (contents + addr, contents + addr + count,
|
||
(size_t) (toaddr - addr - count));
|
||
|
||
/* If we don't have an alignment marker to worry about, we can just
|
||
shrink the section. Otherwise, we have to fill in the newly
|
||
created gap with NOP insns (0x03). */
|
||
if (force_snip)
|
||
sec->size -= count;
|
||
else
|
||
memset (contents + toaddr - count, 0x03, count);
|
||
|
||
/* Adjust all the relocs. */
|
||
for (irel = elf_section_data (sec)->relocs; irel < irelend; irel++)
|
||
{
|
||
/* Get the new reloc address. */
|
||
if (irel->r_offset > addr
|
||
&& (irel->r_offset < toaddr
|
||
|| (force_snip && irel->r_offset == toaddr)))
|
||
irel->r_offset -= count;
|
||
|
||
/* If we see an ALIGN marker at the end of the gap, we move it
|
||
to the beginning of the gap, since marking these gaps is what
|
||
they're for. */
|
||
if (irel->r_offset == toaddr
|
||
&& ELF32_R_TYPE (irel->r_info) == R_RX_RH_RELAX
|
||
&& irel->r_addend & RX_RELAXA_ALIGN)
|
||
irel->r_offset -= count;
|
||
}
|
||
|
||
/* Adjust the local symbols defined in this section. */
|
||
symtab_hdr = &elf_tdata (abfd)->symtab_hdr;
|
||
isym = (Elf_Internal_Sym *) symtab_hdr->contents;
|
||
isymend = isym + symtab_hdr->sh_info;
|
||
|
||
for (; isym < isymend; isym++)
|
||
{
|
||
/* If the symbol is in the range of memory we just moved, we
|
||
have to adjust its value. */
|
||
if (isym->st_shndx == sec_shndx
|
||
&& isym->st_value > addr
|
||
&& isym->st_value < toaddr)
|
||
isym->st_value -= count;
|
||
|
||
/* If the symbol *spans* the bytes we just deleted (i.e. it's
|
||
*end* is in the moved bytes but it's *start* isn't), then we
|
||
must adjust its size. */
|
||
if (isym->st_shndx == sec_shndx
|
||
&& isym->st_value < addr
|
||
&& isym->st_value + isym->st_size > addr
|
||
&& isym->st_value + isym->st_size < toaddr)
|
||
isym->st_size -= count;
|
||
}
|
||
|
||
/* Now adjust the global symbols defined in this section. */
|
||
symcount = (symtab_hdr->sh_size / sizeof (Elf32_External_Sym)
|
||
- symtab_hdr->sh_info);
|
||
sym_hashes = elf_sym_hashes (abfd);
|
||
end_hashes = sym_hashes + symcount;
|
||
|
||
for (; sym_hashes < end_hashes; sym_hashes++)
|
||
{
|
||
struct elf_link_hash_entry *sym_hash = *sym_hashes;
|
||
|
||
if ((sym_hash->root.type == bfd_link_hash_defined
|
||
|| sym_hash->root.type == bfd_link_hash_defweak)
|
||
&& sym_hash->root.u.def.section == sec)
|
||
{
|
||
/* As above, adjust the value if needed. */
|
||
if (sym_hash->root.u.def.value > addr
|
||
&& sym_hash->root.u.def.value < toaddr)
|
||
sym_hash->root.u.def.value -= count;
|
||
|
||
/* As above, adjust the size if needed. */
|
||
if (sym_hash->root.u.def.value < addr
|
||
&& sym_hash->root.u.def.value + sym_hash->size > addr
|
||
&& sym_hash->root.u.def.value + sym_hash->size < toaddr)
|
||
sym_hash->size -= count;
|
||
}
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
/* Used to sort relocs by address. If relocs have the same address,
|
||
we maintain their relative order, except that R_RX_RH_RELAX
|
||
alignment relocs must be the first reloc for any given address. */
|
||
|
||
static void
|
||
reloc_bubblesort (Elf_Internal_Rela * r, int count)
|
||
{
|
||
int i;
|
||
bfd_boolean again;
|
||
bfd_boolean swappit;
|
||
|
||
/* This is almost a classic bubblesort. It's the slowest sort, but
|
||
we're taking advantage of the fact that the relocations are
|
||
mostly in order already (the assembler emits them that way) and
|
||
we need relocs with the same address to remain in the same
|
||
relative order. */
|
||
again = TRUE;
|
||
while (again)
|
||
{
|
||
again = FALSE;
|
||
for (i = 0; i < count - 1; i ++)
|
||
{
|
||
if (r[i].r_offset > r[i + 1].r_offset)
|
||
swappit = TRUE;
|
||
else if (r[i].r_offset < r[i + 1].r_offset)
|
||
swappit = FALSE;
|
||
else if (ELF32_R_TYPE (r[i + 1].r_info) == R_RX_RH_RELAX
|
||
&& (r[i + 1].r_addend & RX_RELAXA_ALIGN))
|
||
swappit = TRUE;
|
||
else if (ELF32_R_TYPE (r[i + 1].r_info) == R_RX_RH_RELAX
|
||
&& (r[i + 1].r_addend & RX_RELAXA_ELIGN)
|
||
&& !(ELF32_R_TYPE (r[i].r_info) == R_RX_RH_RELAX
|
||
&& (r[i].r_addend & RX_RELAXA_ALIGN)))
|
||
swappit = TRUE;
|
||
else
|
||
swappit = FALSE;
|
||
|
||
if (swappit)
|
||
{
|
||
Elf_Internal_Rela tmp;
|
||
|
||
tmp = r[i];
|
||
r[i] = r[i + 1];
|
||
r[i + 1] = tmp;
|
||
/* If we do move a reloc back, re-scan to see if it
|
||
needs to be moved even further back. This avoids
|
||
most of the O(n^2) behavior for our cases. */
|
||
if (i > 0)
|
||
i -= 2;
|
||
again = TRUE;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
|
||
#define OFFSET_FOR_RELOC(rel, lrel, scale) \
|
||
rx_offset_for_reloc (abfd, rel + 1, symtab_hdr, shndx_buf, intsyms, \
|
||
lrel, abfd, sec, link_info, scale)
|
||
|
||
static bfd_vma
|
||
rx_offset_for_reloc (bfd * abfd,
|
||
Elf_Internal_Rela * rel,
|
||
Elf_Internal_Shdr * symtab_hdr,
|
||
Elf_External_Sym_Shndx * shndx_buf ATTRIBUTE_UNUSED,
|
||
Elf_Internal_Sym * intsyms,
|
||
Elf_Internal_Rela ** lrel,
|
||
bfd * input_bfd,
|
||
asection * input_section,
|
||
struct bfd_link_info * info,
|
||
int * scale)
|
||
{
|
||
bfd_vma symval;
|
||
bfd_reloc_status_type r;
|
||
|
||
*scale = 1;
|
||
|
||
/* REL is the first of 1..N relocations. We compute the symbol
|
||
value for each relocation, then combine them if needed. LREL
|
||
gets a pointer to the last relocation used. */
|
||
while (1)
|
||
{
|
||
int32_t tmp1, tmp2;
|
||
|
||
/* Get the value of the symbol referred to by the reloc. */
|
||
if (ELF32_R_SYM (rel->r_info) < symtab_hdr->sh_info)
|
||
{
|
||
/* A local symbol. */
|
||
Elf_Internal_Sym *isym;
|
||
asection *ssec;
|
||
|
||
isym = intsyms + ELF32_R_SYM (rel->r_info);
|
||
|
||
if (isym->st_shndx == SHN_UNDEF)
|
||
ssec = bfd_und_section_ptr;
|
||
else if (isym->st_shndx == SHN_ABS)
|
||
ssec = bfd_abs_section_ptr;
|
||
else if (isym->st_shndx == SHN_COMMON)
|
||
ssec = bfd_com_section_ptr;
|
||
else
|
||
ssec = bfd_section_from_elf_index (abfd,
|
||
isym->st_shndx);
|
||
|
||
/* Initial symbol value. */
|
||
symval = isym->st_value;
|
||
|
||
/* GAS may have made this symbol relative to a section, in
|
||
which case, we have to add the addend to find the
|
||
symbol. */
|
||
if (ELF_ST_TYPE (isym->st_info) == STT_SECTION)
|
||
symval += rel->r_addend;
|
||
|
||
if (ssec)
|
||
{
|
||
if ((ssec->flags & SEC_MERGE)
|
||
&& ssec->sec_info_type == ELF_INFO_TYPE_MERGE)
|
||
symval = _bfd_merged_section_offset (abfd, & ssec,
|
||
elf_section_data (ssec)->sec_info,
|
||
symval);
|
||
}
|
||
|
||
/* Now make the offset relative to where the linker is putting it. */
|
||
if (ssec)
|
||
symval +=
|
||
ssec->output_section->vma + ssec->output_offset;
|
||
|
||
symval += rel->r_addend;
|
||
}
|
||
else
|
||
{
|
||
unsigned long indx;
|
||
struct elf_link_hash_entry * h;
|
||
|
||
/* An external symbol. */
|
||
indx = ELF32_R_SYM (rel->r_info) - symtab_hdr->sh_info;
|
||
h = elf_sym_hashes (abfd)[indx];
|
||
BFD_ASSERT (h != NULL);
|
||
|
||
if (h->root.type != bfd_link_hash_defined
|
||
&& h->root.type != bfd_link_hash_defweak)
|
||
{
|
||
/* This appears to be a reference to an undefined
|
||
symbol. Just ignore it--it will be caught by the
|
||
regular reloc processing. */
|
||
if (lrel)
|
||
*lrel = rel;
|
||
return 0;
|
||
}
|
||
|
||
symval = (h->root.u.def.value
|
||
+ h->root.u.def.section->output_section->vma
|
||
+ h->root.u.def.section->output_offset);
|
||
|
||
symval += rel->r_addend;
|
||
}
|
||
|
||
switch (ELF32_R_TYPE (rel->r_info))
|
||
{
|
||
case R_RX_SYM:
|
||
RX_STACK_PUSH (symval);
|
||
break;
|
||
|
||
case R_RX_OPneg:
|
||
RX_STACK_POP (tmp1);
|
||
tmp1 = - tmp1;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPadd:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 += tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPsub:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp2 -= tmp1;
|
||
RX_STACK_PUSH (tmp2);
|
||
break;
|
||
|
||
case R_RX_OPmul:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 *= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPdiv:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 /= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPshla:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 <<= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPshra:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 >>= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPsctsize:
|
||
RX_STACK_PUSH (input_section->size);
|
||
break;
|
||
|
||
case R_RX_OPscttop:
|
||
RX_STACK_PUSH (input_section->output_section->vma);
|
||
break;
|
||
|
||
case R_RX_OPand:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 &= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPor:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 |= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPxor:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 ^= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPnot:
|
||
RX_STACK_POP (tmp1);
|
||
tmp1 = ~ tmp1;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPmod:
|
||
RX_STACK_POP (tmp1);
|
||
RX_STACK_POP (tmp2);
|
||
tmp1 %= tmp2;
|
||
RX_STACK_PUSH (tmp1);
|
||
break;
|
||
|
||
case R_RX_OPromtop:
|
||
RX_STACK_PUSH (get_romstart (&r, info, input_bfd, input_section, rel->r_offset));
|
||
break;
|
||
|
||
case R_RX_OPramtop:
|
||
RX_STACK_PUSH (get_ramstart (&r, info, input_bfd, input_section, rel->r_offset));
|
||
break;
|
||
|
||
case R_RX_DIR16UL:
|
||
case R_RX_DIR8UL:
|
||
case R_RX_ABS16UL:
|
||
case R_RX_ABS8UL:
|
||
if (rx_stack_top)
|
||
RX_STACK_POP (symval);
|
||
if (lrel)
|
||
*lrel = rel;
|
||
*scale = 4;
|
||
return symval;
|
||
|
||
case R_RX_DIR16UW:
|
||
case R_RX_DIR8UW:
|
||
case R_RX_ABS16UW:
|
||
case R_RX_ABS8UW:
|
||
if (rx_stack_top)
|
||
RX_STACK_POP (symval);
|
||
if (lrel)
|
||
*lrel = rel;
|
||
*scale = 2;
|
||
return symval;
|
||
|
||
default:
|
||
if (rx_stack_top)
|
||
RX_STACK_POP (symval);
|
||
if (lrel)
|
||
*lrel = rel;
|
||
return symval;
|
||
}
|
||
|
||
rel ++;
|
||
}
|
||
}
|
||
|
||
static void
|
||
move_reloc (Elf_Internal_Rela * irel, Elf_Internal_Rela * srel, int delta)
|
||
{
|
||
bfd_vma old_offset = srel->r_offset;
|
||
|
||
irel ++;
|
||
while (irel <= srel)
|
||
{
|
||
if (irel->r_offset == old_offset)
|
||
irel->r_offset += delta;
|
||
irel ++;
|
||
}
|
||
}
|
||
|
||
/* Relax one section. */
|
||
|
||
static bfd_boolean
|
||
elf32_rx_relax_section (bfd * abfd,
|
||
asection * sec,
|
||
struct bfd_link_info * link_info,
|
||
bfd_boolean * again,
|
||
bfd_boolean allow_pcrel3)
|
||
{
|
||
Elf_Internal_Shdr * symtab_hdr;
|
||
Elf_Internal_Shdr * shndx_hdr;
|
||
Elf_Internal_Rela * internal_relocs;
|
||
Elf_Internal_Rela * free_relocs = NULL;
|
||
Elf_Internal_Rela * irel;
|
||
Elf_Internal_Rela * srel;
|
||
Elf_Internal_Rela * irelend;
|
||
Elf_Internal_Rela * next_alignment;
|
||
Elf_Internal_Rela * prev_alignment;
|
||
bfd_byte * contents = NULL;
|
||
bfd_byte * free_contents = NULL;
|
||
Elf_Internal_Sym * intsyms = NULL;
|
||
Elf_Internal_Sym * free_intsyms = NULL;
|
||
Elf_External_Sym_Shndx * shndx_buf = NULL;
|
||
bfd_vma pc;
|
||
bfd_vma sec_start;
|
||
bfd_vma symval = 0;
|
||
int pcrel = 0;
|
||
int code = 0;
|
||
int section_alignment_glue;
|
||
/* how much to scale the relocation by - 1, 2, or 4. */
|
||
int scale;
|
||
|
||
/* Assume nothing changes. */
|
||
*again = FALSE;
|
||
|
||
/* We don't have to do anything for a relocatable link, if
|
||
this section does not have relocs, or if this is not a
|
||
code section. */
|
||
if (link_info->relocatable
|
||
|| (sec->flags & SEC_RELOC) == 0
|
||
|| sec->reloc_count == 0
|
||
|| (sec->flags & SEC_CODE) == 0)
|
||
return TRUE;
|
||
|
||
symtab_hdr = &elf_tdata (abfd)->symtab_hdr;
|
||
shndx_hdr = &elf_tdata (abfd)->symtab_shndx_hdr;
|
||
|
||
sec_start = sec->output_section->vma + sec->output_offset;
|
||
|
||
/* Get the section contents. */
|
||
if (elf_section_data (sec)->this_hdr.contents != NULL)
|
||
contents = elf_section_data (sec)->this_hdr.contents;
|
||
/* Go get them off disk. */
|
||
else
|
||
{
|
||
if (! bfd_malloc_and_get_section (abfd, sec, &contents))
|
||
goto error_return;
|
||
elf_section_data (sec)->this_hdr.contents = contents;
|
||
}
|
||
|
||
/* Read this BFD's symbols. */
|
||
/* Get cached copy if it exists. */
|
||
if (symtab_hdr->contents != NULL)
|
||
intsyms = (Elf_Internal_Sym *) symtab_hdr->contents;
|
||
else
|
||
{
|
||
intsyms = bfd_elf_get_elf_syms (abfd, symtab_hdr, symtab_hdr->sh_info, 0, NULL, NULL, NULL);
|
||
symtab_hdr->contents = (bfd_byte *) intsyms;
|
||
}
|
||
|
||
if (shndx_hdr->sh_size != 0)
|
||
{
|
||
bfd_size_type amt;
|
||
|
||
amt = symtab_hdr->sh_info;
|
||
amt *= sizeof (Elf_External_Sym_Shndx);
|
||
shndx_buf = (Elf_External_Sym_Shndx *) bfd_malloc (amt);
|
||
if (shndx_buf == NULL)
|
||
goto error_return;
|
||
if (bfd_seek (abfd, shndx_hdr->sh_offset, SEEK_SET) != 0
|
||
|| bfd_bread ((PTR) shndx_buf, amt, abfd) != amt)
|
||
goto error_return;
|
||
shndx_hdr->contents = (bfd_byte *) shndx_buf;
|
||
}
|
||
|
||
/* Get a copy of the native relocations. */
|
||
internal_relocs = (_bfd_elf_link_read_relocs
|
||
(abfd, sec, (PTR) NULL, (Elf_Internal_Rela *) NULL,
|
||
link_info->keep_memory));
|
||
if (internal_relocs == NULL)
|
||
goto error_return;
|
||
if (! link_info->keep_memory)
|
||
free_relocs = internal_relocs;
|
||
|
||
/* The RL_ relocs must be just before the operand relocs they go
|
||
with, so we must sort them to guarantee this. We use bubblesort
|
||
instead of qsort so we can guarantee that relocs with the same
|
||
address remain in the same relative order. */
|
||
reloc_bubblesort (internal_relocs, sec->reloc_count);
|
||
|
||
/* Walk through them looking for relaxing opportunities. */
|
||
irelend = internal_relocs + sec->reloc_count;
|
||
|
||
/* This will either be NULL or a pointer to the next alignment
|
||
relocation. */
|
||
next_alignment = internal_relocs;
|
||
/* This will be the previous alignment, although at first it points
|
||
to the first real relocation. */
|
||
prev_alignment = internal_relocs;
|
||
|
||
/* We calculate worst case shrinkage caused by alignment directives.
|
||
No fool-proof, but better than either ignoring the problem or
|
||
doing heavy duty analysis of all the alignment markers in all
|
||
input sections. */
|
||
section_alignment_glue = 0;
|
||
for (irel = internal_relocs; irel < irelend; irel++)
|
||
if (ELF32_R_TYPE (irel->r_info) == R_RX_RH_RELAX
|
||
&& irel->r_addend & RX_RELAXA_ALIGN)
|
||
{
|
||
int this_glue = 1 << (irel->r_addend & RX_RELAXA_ANUM);
|
||
|
||
if (section_alignment_glue < this_glue)
|
||
section_alignment_glue = this_glue;
|
||
}
|
||
/* Worst case is all 0..N alignments, in order, causing 2*N-1 byte
|
||
shrinkage. */
|
||
section_alignment_glue *= 2;
|
||
|
||
for (irel = internal_relocs; irel < irelend; irel++)
|
||
{
|
||
unsigned char *insn;
|
||
int nrelocs;
|
||
|
||
/* The insns we care about are all marked with one of these. */
|
||
if (ELF32_R_TYPE (irel->r_info) != R_RX_RH_RELAX)
|
||
continue;
|
||
|
||
if (irel->r_addend & RX_RELAXA_ALIGN
|
||
|| next_alignment == internal_relocs)
|
||
{
|
||
/* When we delete bytes, we need to maintain all the alignments
|
||
indicated. In addition, we need to be careful about relaxing
|
||
jumps across alignment boundaries - these displacements
|
||
*grow* when we delete bytes. For now, don't shrink
|
||
displacements across an alignment boundary, just in case.
|
||
Note that this only affects relocations to the same
|
||
section. */
|
||
prev_alignment = next_alignment;
|
||
next_alignment += 2;
|
||
while (next_alignment < irelend
|
||
&& (ELF32_R_TYPE (next_alignment->r_info) != R_RX_RH_RELAX
|
||
|| !(next_alignment->r_addend & RX_RELAXA_ELIGN)))
|
||
next_alignment ++;
|
||
if (next_alignment >= irelend || next_alignment->r_offset == 0)
|
||
next_alignment = NULL;
|
||
}
|
||
|
||
/* When we hit alignment markers, see if we've shrunk enough
|
||
before them to reduce the gap without violating the alignment
|
||
requirements. */
|
||
if (irel->r_addend & RX_RELAXA_ALIGN)
|
||
{
|
||
/* At this point, the next relocation *should* be the ELIGN
|
||
end marker. */
|
||
Elf_Internal_Rela *erel = irel + 1;
|
||
unsigned int alignment, nbytes;
|
||
|
||
if (ELF32_R_TYPE (erel->r_info) != R_RX_RH_RELAX)
|
||
continue;
|
||
if (!(erel->r_addend & RX_RELAXA_ELIGN))
|
||
continue;
|
||
|
||
alignment = 1 << (irel->r_addend & RX_RELAXA_ANUM);
|
||
|
||
if (erel->r_offset - irel->r_offset < alignment)
|
||
continue;
|
||
|
||
nbytes = erel->r_offset - irel->r_offset;
|
||
nbytes /= alignment;
|
||
nbytes *= alignment;
|
||
|
||
elf32_rx_relax_delete_bytes (abfd, sec, erel->r_offset-nbytes, nbytes, next_alignment,
|
||
erel->r_offset == sec->size);
|
||
*again = TRUE;
|
||
|
||
continue;
|
||
}
|
||
|
||
if (irel->r_addend & RX_RELAXA_ELIGN)
|
||
continue;
|
||
|
||
insn = contents + irel->r_offset;
|
||
|
||
nrelocs = irel->r_addend & RX_RELAXA_RNUM;
|
||
|
||
/* At this point, we have an insn that is a candidate for linker
|
||
relaxation. There are NRELOCS relocs following that may be
|
||
relaxed, although each reloc may be made of more than one
|
||
reloc entry (such as gp-rel symbols). */
|
||
|
||
/* Get the value of the symbol referred to by the reloc. Just
|
||
in case this is the last reloc in the list, use the RL's
|
||
addend to choose between this reloc (no addend) or the next
|
||
(yes addend, which means at least one following reloc). */
|
||
|
||
/* srel points to the "current" reloction for this insn -
|
||
actually the last reloc for a given operand, which is the one
|
||
we need to update. We check the relaxations in the same
|
||
order that the relocations happen, so we'll just push it
|
||
along as we go. */
|
||
srel = irel;
|
||
|
||
pc = sec->output_section->vma + sec->output_offset
|
||
+ srel->r_offset;
|
||
|
||
#define GET_RELOC \
|
||
symval = OFFSET_FOR_RELOC (srel, &srel, &scale); \
|
||
pcrel = symval - pc + srel->r_addend; \
|
||
nrelocs --;
|
||
|
||
#define SNIPNR(offset, nbytes) \
|
||
elf32_rx_relax_delete_bytes (abfd, sec, (insn - contents) + offset, nbytes, next_alignment, 0);
|
||
#define SNIP(offset, nbytes, newtype) \
|
||
SNIPNR (offset, nbytes); \
|
||
srel->r_info = ELF32_R_INFO (ELF32_R_SYM (srel->r_info), newtype)
|
||
|
||
/* The order of these bit tests must match the order that the
|
||
relocs appear in. Since we sorted those by offset, we can
|
||
predict them. */
|
||
|
||
/* Note that the numbers in, say, DSP6 are the bit offsets of
|
||
the code fields that describe the operand. Bits number 0 for
|
||
the MSB of insn[0]. */
|
||
|
||
/* DSP* codes:
|
||
0 00 [reg]
|
||
1 01 dsp:8[reg]
|
||
2 10 dsp:16[reg]
|
||
3 11 reg */
|
||
if (irel->r_addend & RX_RELAXA_DSP6)
|
||
{
|
||
GET_RELOC;
|
||
|
||
code = insn[0] & 3;
|
||
if (code == 2 && symval/scale <= 255)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
insn[0] &= 0xfc;
|
||
insn[0] |= 0x01;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (3, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
else if (code == 1 && symval == 0)
|
||
{
|
||
insn[0] &= 0xfc;
|
||
SNIP (2, 1, R_RX_NONE);
|
||
*again = TRUE;
|
||
}
|
||
|
||
/* Special case DSP:5 format: MOV.bwl dsp:5[Rsrc],Rdst. */
|
||
else if (code == 1 && symval/scale <= 31
|
||
/* Decodable bits. */
|
||
&& (insn[0] & 0xcc) == 0xcc
|
||
/* Width. */
|
||
&& (insn[0] & 0x30) != 0x30
|
||
/* Register MSBs. */
|
||
&& (insn[1] & 0x88) == 0x00)
|
||
{
|
||
int newrel = 0;
|
||
|
||
insn[0] = 0x88 | (insn[0] & 0x30);
|
||
/* The register fields are in the right place already. */
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
switch ((insn[0] & 0x30) >> 4)
|
||
{
|
||
case 0:
|
||
newrel = R_RX_RH_ABS5p5B;
|
||
break;
|
||
case 1:
|
||
newrel = R_RX_RH_ABS5p5W;
|
||
break;
|
||
case 2:
|
||
newrel = R_RX_RH_ABS5p5L;
|
||
break;
|
||
}
|
||
|
||
move_reloc (irel, srel, -2);
|
||
SNIP (2, 1, newrel);
|
||
}
|
||
|
||
/* Special case DSP:5 format: MOVU.bw dsp:5[Rsrc],Rdst. */
|
||
else if (code == 1 && symval/scale <= 31
|
||
/* Decodable bits. */
|
||
&& (insn[0] & 0xf8) == 0x58
|
||
/* Register MSBs. */
|
||
&& (insn[1] & 0x88) == 0x00)
|
||
{
|
||
int newrel = 0;
|
||
|
||
insn[0] = 0xb0 | ((insn[0] & 0x04) << 1);
|
||
/* The register fields are in the right place already. */
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
switch ((insn[0] & 0x08) >> 3)
|
||
{
|
||
case 0:
|
||
newrel = R_RX_RH_ABS5p5B;
|
||
break;
|
||
case 1:
|
||
newrel = R_RX_RH_ABS5p5W;
|
||
break;
|
||
}
|
||
|
||
move_reloc (irel, srel, -2);
|
||
SNIP (2, 1, newrel);
|
||
}
|
||
}
|
||
|
||
/* A DSP4 operand always follows a DSP6 operand, even if there's
|
||
no relocation for it. We have to read the code out of the
|
||
opcode to calculate the offset of the operand. */
|
||
if (irel->r_addend & RX_RELAXA_DSP4)
|
||
{
|
||
int code6, offset = 0;
|
||
|
||
GET_RELOC;
|
||
|
||
code6 = insn[0] & 0x03;
|
||
switch (code6)
|
||
{
|
||
case 0: offset = 2; break;
|
||
case 1: offset = 3; break;
|
||
case 2: offset = 4; break;
|
||
case 3: offset = 2; break;
|
||
}
|
||
|
||
code = (insn[0] & 0x0c) >> 2;
|
||
|
||
if (code == 2 && symval / scale <= 255)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[0] &= 0xf3;
|
||
insn[0] |= 0x04;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (offset+1, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
else if (code == 1 && symval == 0)
|
||
{
|
||
insn[0] &= 0xf3;
|
||
SNIP (offset, 1, R_RX_NONE);
|
||
*again = TRUE;
|
||
}
|
||
/* Special case DSP:5 format: MOV.bwl Rsrc,dsp:5[Rdst] */
|
||
else if (code == 1 && symval/scale <= 31
|
||
/* Decodable bits. */
|
||
&& (insn[0] & 0xc3) == 0xc3
|
||
/* Width. */
|
||
&& (insn[0] & 0x30) != 0x30
|
||
/* Register MSBs. */
|
||
&& (insn[1] & 0x88) == 0x00)
|
||
{
|
||
int newrel = 0;
|
||
|
||
insn[0] = 0x80 | (insn[0] & 0x30);
|
||
/* The register fields are in the right place already. */
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
switch ((insn[0] & 0x30) >> 4)
|
||
{
|
||
case 0:
|
||
newrel = R_RX_RH_ABS5p5B;
|
||
break;
|
||
case 1:
|
||
newrel = R_RX_RH_ABS5p5W;
|
||
break;
|
||
case 2:
|
||
newrel = R_RX_RH_ABS5p5L;
|
||
break;
|
||
}
|
||
|
||
move_reloc (irel, srel, -2);
|
||
SNIP (2, 1, newrel);
|
||
}
|
||
}
|
||
|
||
/* These always occur alone, but the offset depends on whether
|
||
it's a MEMEX opcode (0x06) or not. */
|
||
if (irel->r_addend & RX_RELAXA_DSP14)
|
||
{
|
||
int offset;
|
||
GET_RELOC;
|
||
|
||
if (insn[0] == 0x06)
|
||
offset = 3;
|
||
else
|
||
offset = 4;
|
||
|
||
code = insn[1] & 3;
|
||
|
||
if (code == 2 && symval / scale <= 255)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[1] &= 0xfc;
|
||
insn[1] |= 0x01;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (offset, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
else if (code == 1 && symval == 0)
|
||
{
|
||
insn[1] &= 0xfc;
|
||
SNIP (offset, 1, R_RX_NONE);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* IMM* codes:
|
||
0 00 imm:32
|
||
1 01 simm:8
|
||
2 10 simm:16
|
||
3 11 simm:24. */
|
||
|
||
/* These always occur alone. */
|
||
if (irel->r_addend & RX_RELAXA_IMM6)
|
||
{
|
||
long ssymval;
|
||
|
||
GET_RELOC;
|
||
|
||
/* These relocations sign-extend, so we must do signed compares. */
|
||
ssymval = (long) symval;
|
||
|
||
code = insn[0] & 0x03;
|
||
|
||
if (code == 0 && ssymval <= 8388607 && ssymval >= -8388608)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[0] &= 0xfc;
|
||
insn[0] |= 0x03;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
else if (code == 3 && ssymval <= 32767 && ssymval >= -32768)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[0] &= 0xfc;
|
||
insn[0] |= 0x02;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* Special case UIMM8 format: CMP #uimm8,Rdst. */
|
||
else if (code == 2 && ssymval <= 255 && ssymval >= 16
|
||
/* Decodable bits. */
|
||
&& (insn[0] & 0xfc) == 0x74
|
||
/* Decodable bits. */
|
||
&& ((insn[1] & 0xf0) == 0x00))
|
||
{
|
||
int newrel;
|
||
|
||
insn[0] = 0x75;
|
||
insn[1] = 0x50 | (insn[1] & 0x0f);
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
if (STACK_REL_P (ELF32_R_TYPE (srel->r_info)))
|
||
newrel = R_RX_ABS8U;
|
||
else
|
||
newrel = R_RX_DIR8U;
|
||
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
|
||
else if (code == 2 && ssymval <= 127 && ssymval >= -128)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[0] &= 0xfc;
|
||
insn[0] |= 0x01;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* Special case UIMM4 format: CMP, MUL, AND, OR. */
|
||
else if (code == 1 && ssymval <= 15 && ssymval >= 0
|
||
/* Decodable bits and immediate type. */
|
||
&& insn[0] == 0x75
|
||
/* Decodable bits. */
|
||
&& (insn[1] & 0xc0) == 0x00)
|
||
{
|
||
static const int newop[4] = { 1, 3, 4, 5 };
|
||
|
||
insn[0] = 0x60 | newop[insn[1] >> 4];
|
||
/* The register number doesn't move. */
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
move_reloc (irel, srel, -1);
|
||
|
||
SNIP (2, 1, R_RX_RH_UIMM4p8);
|
||
*again = TRUE;
|
||
}
|
||
|
||
/* Special case UIMM4 format: ADD -> ADD/SUB. */
|
||
else if (code == 1 && ssymval <= 15 && ssymval >= -15
|
||
/* Decodable bits and immediate type. */
|
||
&& insn[0] == 0x71
|
||
/* Same register for source and destination. */
|
||
&& ((insn[1] >> 4) == (insn[1] & 0x0f)))
|
||
{
|
||
int newrel;
|
||
|
||
/* Note that we can't turn "add $0,Rs" into a NOP
|
||
because the flags need to be set right. */
|
||
|
||
if (ssymval < 0)
|
||
{
|
||
insn[0] = 0x60; /* Subtract. */
|
||
newrel = R_RX_RH_UNEG4p8;
|
||
}
|
||
else
|
||
{
|
||
insn[0] = 0x62; /* Add. */
|
||
newrel = R_RX_RH_UIMM4p8;
|
||
}
|
||
|
||
/* The register number is in the right place. */
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
move_reloc (irel, srel, -1);
|
||
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* These are either matched with a DSP6 (2-byte base) or an id24
|
||
(3-byte base). */
|
||
if (irel->r_addend & RX_RELAXA_IMM12)
|
||
{
|
||
int dspcode, offset = 0;
|
||
long ssymval;
|
||
|
||
GET_RELOC;
|
||
|
||
if ((insn[0] & 0xfc) == 0xfc)
|
||
dspcode = 1; /* Just something with one byte operand. */
|
||
else
|
||
dspcode = insn[0] & 3;
|
||
switch (dspcode)
|
||
{
|
||
case 0: offset = 2; break;
|
||
case 1: offset = 3; break;
|
||
case 2: offset = 4; break;
|
||
case 3: offset = 2; break;
|
||
}
|
||
|
||
/* These relocations sign-extend, so we must do signed compares. */
|
||
ssymval = (long) symval;
|
||
|
||
code = (insn[1] >> 2) & 3;
|
||
if (code == 0 && ssymval <= 8388607 && ssymval >= -8388608)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[1] &= 0xf3;
|
||
insn[1] |= 0x0c;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (offset, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
else if (code == 3 && ssymval <= 32767 && ssymval >= -32768)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
|
||
insn[1] &= 0xf3;
|
||
insn[1] |= 0x08;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE (srel->r_info))
|
||
{
|
||
SNIP (offset, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* Special case UIMM8 format: MOV #uimm8,Rdst. */
|
||
else if (code == 2 && ssymval <= 255 && ssymval >= 16
|
||
/* Decodable bits. */
|
||
&& insn[0] == 0xfb
|
||
/* Decodable bits. */
|
||
&& ((insn[1] & 0x03) == 0x02))
|
||
{
|
||
int newrel;
|
||
|
||
insn[0] = 0x75;
|
||
insn[1] = 0x40 | (insn[1] >> 4);
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
if (STACK_REL_P (ELF32_R_TYPE (srel->r_info)))
|
||
newrel = R_RX_ABS8U;
|
||
else
|
||
newrel = R_RX_DIR8U;
|
||
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
|
||
else if (code == 2 && ssymval <= 127 && ssymval >= -128)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE(srel->r_info);
|
||
|
||
insn[1] &= 0xf3;
|
||
insn[1] |= 0x04;
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
if (newrel != ELF32_R_TYPE(srel->r_info))
|
||
{
|
||
SNIP (offset, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
/* Special case UIMM4 format: MOV #uimm4,Rdst. */
|
||
else if (code == 1 && ssymval <= 15 && ssymval >= 0
|
||
/* Decodable bits. */
|
||
&& insn[0] == 0xfb
|
||
/* Decodable bits. */
|
||
&& ((insn[1] & 0x03) == 0x02))
|
||
{
|
||
insn[0] = 0x66;
|
||
insn[1] = insn[1] >> 4;
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
|
||
move_reloc (irel, srel, -1);
|
||
|
||
SNIP (2, 1, R_RX_RH_UIMM4p8);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
if (irel->r_addend & RX_RELAXA_BRA)
|
||
{
|
||
unsigned int newrel = ELF32_R_TYPE (srel->r_info);
|
||
int max_pcrel3 = 4;
|
||
int alignment_glue = 0;
|
||
|
||
GET_RELOC;
|
||
|
||
/* Branches over alignment chunks are problematic, as
|
||
deleting bytes here makes the branch *further* away. We
|
||
can be agressive with branches within this alignment
|
||
block, but not branches outside it. */
|
||
if ((prev_alignment == NULL
|
||
|| symval < (bfd_vma)(sec_start + prev_alignment->r_offset))
|
||
&& (next_alignment == NULL
|
||
|| symval > (bfd_vma)(sec_start + next_alignment->r_offset)))
|
||
alignment_glue = section_alignment_glue;
|
||
|
||
if (ELF32_R_TYPE(srel[1].r_info) == R_RX_RH_RELAX
|
||
&& srel[1].r_addend & RX_RELAXA_BRA
|
||
&& srel[1].r_offset < irel->r_offset + pcrel)
|
||
max_pcrel3 ++;
|
||
|
||
newrel = next_smaller_reloc (ELF32_R_TYPE (srel->r_info));
|
||
|
||
/* The values we compare PCREL with are not what you'd
|
||
expect; they're off by a little to compensate for (1)
|
||
where the reloc is relative to the insn, and (2) how much
|
||
the insn is going to change when we relax it. */
|
||
|
||
/* These we have to decode. */
|
||
switch (insn[0])
|
||
{
|
||
case 0x04: /* BRA pcdsp:24 */
|
||
if (-32768 + alignment_glue <= pcrel
|
||
&& pcrel <= 32765 - alignment_glue)
|
||
{
|
||
insn[0] = 0x38;
|
||
SNIP (3, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x38: /* BRA pcdsp:16 */
|
||
if (-128 + alignment_glue <= pcrel
|
||
&& pcrel <= 127 - alignment_glue)
|
||
{
|
||
insn[0] = 0x2e;
|
||
SNIP (2, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x2e: /* BRA pcdsp:8 */
|
||
/* Note that there's a risk here of shortening things so
|
||
much that we no longer fit this reloc; it *should*
|
||
only happen when you branch across a branch, and that
|
||
branch also devolves into BRA.S. "Real" code should
|
||
be OK. */
|
||
if (max_pcrel3 + alignment_glue <= pcrel
|
||
&& pcrel <= 10 - alignment_glue
|
||
&& allow_pcrel3)
|
||
{
|
||
insn[0] = 0x08;
|
||
SNIP (1, 1, newrel);
|
||
move_reloc (irel, srel, -1);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x05: /* BSR pcdsp:24 */
|
||
if (-32768 + alignment_glue <= pcrel
|
||
&& pcrel <= 32765 - alignment_glue)
|
||
{
|
||
insn[0] = 0x39;
|
||
SNIP (1, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x3a: /* BEQ.W pcdsp:16 */
|
||
case 0x3b: /* BNE.W pcdsp:16 */
|
||
if (-128 + alignment_glue <= pcrel
|
||
&& pcrel <= 127 - alignment_glue)
|
||
{
|
||
insn[0] = 0x20 | (insn[0] & 1);
|
||
SNIP (1, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x20: /* BEQ.B pcdsp:8 */
|
||
case 0x21: /* BNE.B pcdsp:8 */
|
||
if (max_pcrel3 + alignment_glue <= pcrel
|
||
&& pcrel - alignment_glue <= 10
|
||
&& allow_pcrel3)
|
||
{
|
||
insn[0] = 0x10 | ((insn[0] & 1) << 3);
|
||
SNIP (1, 1, newrel);
|
||
move_reloc (irel, srel, -1);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
|
||
case 0x16: /* synthetic BNE dsp24 */
|
||
case 0x1e: /* synthetic BEQ dsp24 */
|
||
if (-32767 + alignment_glue <= pcrel
|
||
&& pcrel <= 32766 - alignment_glue
|
||
&& insn[1] == 0x04)
|
||
{
|
||
if (insn[0] == 0x16)
|
||
insn[0] = 0x3b;
|
||
else
|
||
insn[0] = 0x3a;
|
||
/* We snip out the bytes at the end else the reloc
|
||
will get moved too, and too much. */
|
||
SNIP (3, 2, newrel);
|
||
move_reloc (irel, srel, -1);
|
||
*again = TRUE;
|
||
}
|
||
break;
|
||
}
|
||
|
||
/* Special case - synthetic conditional branches, pcrel24.
|
||
Note that EQ and NE have been handled above. */
|
||
if ((insn[0] & 0xf0) == 0x20
|
||
&& insn[1] == 0x06
|
||
&& insn[2] == 0x04
|
||
&& srel->r_offset != irel->r_offset + 1
|
||
&& -32767 + alignment_glue <= pcrel
|
||
&& pcrel <= 32766 - alignment_glue)
|
||
{
|
||
insn[1] = 0x05;
|
||
insn[2] = 0x38;
|
||
SNIP (5, 1, newrel);
|
||
*again = TRUE;
|
||
}
|
||
|
||
/* Special case - synthetic conditional branches, pcrel16 */
|
||
if ((insn[0] & 0xf0) == 0x20
|
||
&& insn[1] == 0x05
|
||
&& insn[2] == 0x38
|
||
&& srel->r_offset != irel->r_offset + 1
|
||
&& -127 + alignment_glue <= pcrel
|
||
&& pcrel <= 126 - alignment_glue)
|
||
{
|
||
int cond = (insn[0] & 0x0f) ^ 0x01;
|
||
|
||
insn[0] = 0x20 | cond;
|
||
/* By moving the reloc first, we avoid having
|
||
delete_bytes move it also. */
|
||
move_reloc (irel, srel, -2);
|
||
SNIP (2, 3, newrel);
|
||
*again = TRUE;
|
||
}
|
||
}
|
||
|
||
BFD_ASSERT (nrelocs == 0);
|
||
|
||
/* Special case - check MOV.bwl #IMM, dsp[reg] and see if we can
|
||
use MOV.bwl #uimm:8, dsp:5[r7] format. This is tricky
|
||
because it may have one or two relocations. */
|
||
if ((insn[0] & 0xfc) == 0xf8
|
||
&& (insn[1] & 0x80) == 0x00
|
||
&& (insn[0] & 0x03) != 0x03)
|
||
{
|
||
int dcode, icode, reg, ioff, dscale, ilen;
|
||
bfd_vma disp_val = 0;
|
||
long imm_val = 0;
|
||
Elf_Internal_Rela * disp_rel = 0;
|
||
Elf_Internal_Rela * imm_rel = 0;
|
||
|
||
/* Reset this. */
|
||
srel = irel;
|
||
|
||
dcode = insn[0] & 0x03;
|
||
icode = (insn[1] >> 2) & 0x03;
|
||
reg = (insn[1] >> 4) & 0x0f;
|
||
|
||
ioff = dcode == 1 ? 3 : dcode == 2 ? 4 : 2;
|
||
|
||
/* Figure out what the dispacement is. */
|
||
if (dcode == 1 || dcode == 2)
|
||
{
|
||
/* There's a displacement. See if there's a reloc for it. */
|
||
if (srel[1].r_offset == irel->r_offset + 2)
|
||
{
|
||
GET_RELOC;
|
||
disp_val = symval;
|
||
disp_rel = srel;
|
||
}
|
||
else
|
||
{
|
||
if (dcode == 1)
|
||
disp_val = insn[2];
|
||
else
|
||
{
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
disp_val = insn[2] * 256 + insn[3];
|
||
#else
|
||
disp_val = insn[2] + insn[3] * 256;
|
||
#endif
|
||
}
|
||
switch (insn[1] & 3)
|
||
{
|
||
case 1:
|
||
disp_val *= 2;
|
||
scale = 2;
|
||
break;
|
||
case 2:
|
||
disp_val *= 4;
|
||
scale = 4;
|
||
break;
|
||
}
|
||
}
|
||
}
|
||
|
||
dscale = scale;
|
||
|
||
/* Figure out what the immediate is. */
|
||
if (srel[1].r_offset == irel->r_offset + ioff)
|
||
{
|
||
GET_RELOC;
|
||
imm_val = (long) symval;
|
||
imm_rel = srel;
|
||
}
|
||
else
|
||
{
|
||
unsigned char * ip = insn + ioff;
|
||
|
||
switch (icode)
|
||
{
|
||
case 1:
|
||
/* For byte writes, we don't sign extend. Makes the math easier later. */
|
||
if (scale == 1)
|
||
imm_val = ip[0];
|
||
else
|
||
imm_val = (char) ip[0];
|
||
break;
|
||
case 2:
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
imm_val = ((char) ip[0] << 8) | ip[1];
|
||
#else
|
||
imm_val = ((char) ip[1] << 8) | ip[0];
|
||
#endif
|
||
break;
|
||
case 3:
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
imm_val = ((char) ip[0] << 16) | (ip[1] << 8) | ip[2];
|
||
#else
|
||
imm_val = ((char) ip[2] << 16) | (ip[1] << 8) | ip[0];
|
||
#endif
|
||
break;
|
||
case 0:
|
||
#if RX_OPCODE_BIG_ENDIAN
|
||
imm_val = (ip[0] << 24) | (ip[1] << 16) | (ip[2] << 8) | ip[3];
|
||
#else
|
||
imm_val = (ip[3] << 24) | (ip[2] << 16) | (ip[1] << 8) | ip[0];
|
||
#endif
|
||
break;
|
||
}
|
||
}
|
||
|
||
ilen = 2;
|
||
|
||
switch (dcode)
|
||
{
|
||
case 1:
|
||
ilen += 1;
|
||
break;
|
||
case 2:
|
||
ilen += 2;
|
||
break;
|
||
}
|
||
|
||
switch (icode)
|
||
{
|
||
case 1:
|
||
ilen += 1;
|
||
break;
|
||
case 2:
|
||
ilen += 2;
|
||
break;
|
||
case 3:
|
||
ilen += 3;
|
||
break;
|
||
case 4:
|
||
ilen += 4;
|
||
break;
|
||
}
|
||
|
||
/* The shortcut happens when the immediate is 0..255,
|
||
register r0 to r7, and displacement (scaled) 0..31. */
|
||
|
||
if (0 <= imm_val && imm_val <= 255
|
||
&& 0 <= reg && reg <= 7
|
||
&& disp_val / dscale <= 31)
|
||
{
|
||
insn[0] = 0x3c | (insn[1] & 0x03);
|
||
insn[1] = (((disp_val / dscale) << 3) & 0x80) | (reg << 4) | ((disp_val/dscale) & 0x0f);
|
||
insn[2] = imm_val;
|
||
|
||
if (disp_rel)
|
||
{
|
||
int newrel = R_RX_NONE;
|
||
|
||
switch (dscale)
|
||
{
|
||
case 1:
|
||
newrel = R_RX_RH_ABS5p8B;
|
||
break;
|
||
case 2:
|
||
newrel = R_RX_RH_ABS5p8W;
|
||
break;
|
||
case 4:
|
||
newrel = R_RX_RH_ABS5p8L;
|
||
break;
|
||
}
|
||
disp_rel->r_info = ELF32_R_INFO (ELF32_R_SYM (disp_rel->r_info), newrel);
|
||
move_reloc (irel, disp_rel, -1);
|
||
}
|
||
if (imm_rel)
|
||
{
|
||
imm_rel->r_info = ELF32_R_INFO (ELF32_R_SYM (imm_rel->r_info), R_RX_DIR8U);
|
||
move_reloc (disp_rel ? disp_rel : irel,
|
||
imm_rel,
|
||
irel->r_offset - imm_rel->r_offset + 2);
|
||
}
|
||
|
||
SNIPNR (3, ilen - 3);
|
||
*again = TRUE;
|
||
|
||
/* We can't relax this new opcode. */
|
||
irel->r_addend = 0;
|
||
}
|
||
}
|
||
}
|
||
|
||
/* We can't reliably relax branches to DIR3U_PCREL unless we know
|
||
whatever they're branching over won't shrink any more. If we're
|
||
basically done here, do one more pass just for branches - but
|
||
don't request a pass after that one! */
|
||
if (!*again && !allow_pcrel3)
|
||
{
|
||
bfd_boolean ignored;
|
||
|
||
elf32_rx_relax_section (abfd, sec, link_info, &ignored, TRUE);
|
||
}
|
||
|
||
return TRUE;
|
||
|
||
error_return:
|
||
if (free_relocs != NULL)
|
||
free (free_relocs);
|
||
|
||
if (free_contents != NULL)
|
||
free (free_contents);
|
||
|
||
if (shndx_buf != NULL)
|
||
{
|
||
shndx_hdr->contents = NULL;
|
||
free (shndx_buf);
|
||
}
|
||
|
||
if (free_intsyms != NULL)
|
||
free (free_intsyms);
|
||
|
||
return FALSE;
|
||
}
|
||
|
||
static bfd_boolean
|
||
elf32_rx_relax_section_wrapper (bfd * abfd,
|
||
asection * sec,
|
||
struct bfd_link_info * link_info,
|
||
bfd_boolean * again)
|
||
{
|
||
return elf32_rx_relax_section (abfd, sec, link_info, again, FALSE);
|
||
}
|
||
|
||
/* Function to set the ELF flag bits. */
|
||
|
||
static bfd_boolean
|
||
rx_elf_set_private_flags (bfd * abfd, flagword flags)
|
||
{
|
||
elf_elfheader (abfd)->e_flags = flags;
|
||
elf_flags_init (abfd) = TRUE;
|
||
return TRUE;
|
||
}
|
||
|
||
static bfd_boolean no_warn_mismatch = FALSE;
|
||
static bfd_boolean ignore_lma = TRUE;
|
||
|
||
void bfd_elf32_rx_set_target_flags (bfd_boolean, bfd_boolean);
|
||
|
||
void
|
||
bfd_elf32_rx_set_target_flags (bfd_boolean user_no_warn_mismatch,
|
||
bfd_boolean user_ignore_lma)
|
||
{
|
||
no_warn_mismatch = user_no_warn_mismatch;
|
||
ignore_lma = user_ignore_lma;
|
||
}
|
||
|
||
/* Merge backend specific data from an object file to the output
|
||
object file when linking. */
|
||
|
||
static bfd_boolean
|
||
rx_elf_merge_private_bfd_data (bfd * ibfd, bfd * obfd)
|
||
{
|
||
flagword old_flags;
|
||
flagword new_flags;
|
||
bfd_boolean error = FALSE;
|
||
|
||
new_flags = elf_elfheader (ibfd)->e_flags;
|
||
old_flags = elf_elfheader (obfd)->e_flags;
|
||
|
||
if (!elf_flags_init (obfd))
|
||
{
|
||
/* First call, no flags set. */
|
||
elf_flags_init (obfd) = TRUE;
|
||
elf_elfheader (obfd)->e_flags = new_flags;
|
||
}
|
||
else if (old_flags != new_flags)
|
||
{
|
||
flagword known_flags = E_FLAG_RX_64BIT_DOUBLES | E_FLAG_RX_DSP | E_FLAG_RX_PID;
|
||
|
||
if ((old_flags ^ new_flags) & known_flags)
|
||
{
|
||
/* Only complain if flag bits we care about do not match.
|
||
Other bits may be set, since older binaries did use some
|
||
deprecated flags. */
|
||
if (no_warn_mismatch)
|
||
{
|
||
elf_elfheader (obfd)->e_flags = (new_flags | old_flags) & known_flags;
|
||
}
|
||
else
|
||
{
|
||
(*_bfd_error_handler)
|
||
("ELF header flags mismatch: old_flags = 0x%.8lx, new_flags = 0x%.8lx, filename = %s",
|
||
old_flags, new_flags, bfd_get_filename (ibfd));
|
||
error = TRUE;
|
||
}
|
||
}
|
||
else
|
||
elf_elfheader (obfd)->e_flags = new_flags & known_flags;
|
||
}
|
||
|
||
if (error)
|
||
bfd_set_error (bfd_error_bad_value);
|
||
|
||
return !error;
|
||
}
|
||
|
||
static bfd_boolean
|
||
rx_elf_print_private_bfd_data (bfd * abfd, void * ptr)
|
||
{
|
||
FILE * file = (FILE *) ptr;
|
||
flagword flags;
|
||
|
||
BFD_ASSERT (abfd != NULL && ptr != NULL);
|
||
|
||
/* Print normal ELF private data. */
|
||
_bfd_elf_print_private_bfd_data (abfd, ptr);
|
||
|
||
flags = elf_elfheader (abfd)->e_flags;
|
||
fprintf (file, _("private flags = 0x%lx:"), (long) flags);
|
||
|
||
if (flags & E_FLAG_RX_64BIT_DOUBLES)
|
||
fprintf (file, _(" [64-bit doubles]"));
|
||
if (flags & E_FLAG_RX_DSP)
|
||
fprintf (file, _(" [dsp]"));
|
||
|
||
fputc ('\n', file);
|
||
return TRUE;
|
||
}
|
||
|
||
/* Return the MACH for an e_flags value. */
|
||
|
||
static int
|
||
elf32_rx_machine (bfd * abfd)
|
||
{
|
||
if ((elf_elfheader (abfd)->e_flags & EF_RX_CPU_MASK) == EF_RX_CPU_RX)
|
||
return bfd_mach_rx;
|
||
|
||
return 0;
|
||
}
|
||
|
||
static bfd_boolean
|
||
rx_elf_object_p (bfd * abfd)
|
||
{
|
||
int i;
|
||
unsigned int u;
|
||
Elf_Internal_Phdr *phdr = elf_tdata (abfd)->phdr;
|
||
int nphdrs = elf_elfheader (abfd)->e_phnum;
|
||
sec_ptr bsec;
|
||
static int saw_be = FALSE;
|
||
|
||
/* We never want to automatically choose the non-swapping big-endian
|
||
target. The user can only get that explicitly, such as with -I
|
||
and objcopy. */
|
||
if (abfd->xvec == &bfd_elf32_rx_be_ns_vec
|
||
&& abfd->target_defaulted)
|
||
return FALSE;
|
||
|
||
/* BFD->target_defaulted is not set to TRUE when a target is chosen
|
||
as a fallback, so we check for "scanning" to know when to stop
|
||
using the non-swapping target. */
|
||
if (abfd->xvec == &bfd_elf32_rx_be_ns_vec
|
||
&& saw_be)
|
||
return FALSE;
|
||
if (abfd->xvec == &bfd_elf32_rx_be_vec)
|
||
saw_be = TRUE;
|
||
|
||
bfd_default_set_arch_mach (abfd, bfd_arch_rx,
|
||
elf32_rx_machine (abfd));
|
||
|
||
/* For each PHDR in the object, we must find some section that
|
||
corresponds (based on matching file offsets) and use its VMA
|
||
information to reconstruct the p_vaddr field we clobbered when we
|
||
wrote it out. */
|
||
for (i=0; i<nphdrs; i++)
|
||
{
|
||
for (u=0; u<elf_tdata(abfd)->num_elf_sections; u++)
|
||
{
|
||
Elf_Internal_Shdr *sec = elf_tdata(abfd)->elf_sect_ptr[u];
|
||
|
||
if (phdr[i].p_offset <= (bfd_vma) sec->sh_offset
|
||
&& (bfd_vma)sec->sh_offset <= phdr[i].p_offset + (phdr[i].p_filesz - 1))
|
||
{
|
||
/* Found one! The difference between the two addresses,
|
||
plus the difference between the two file offsets, is
|
||
enough information to reconstruct the lma. */
|
||
|
||
/* Example where they aren't:
|
||
PHDR[1] = lma fffc0100 offset 00002010 size 00000100
|
||
SEC[6] = vma 00000050 offset 00002050 size 00000040
|
||
|
||
The correct LMA for the section is fffc0140 + (2050-2010).
|
||
*/
|
||
|
||
phdr[i].p_vaddr = sec->sh_addr + (sec->sh_offset - phdr[i].p_offset);
|
||
break;
|
||
}
|
||
}
|
||
|
||
/* We must update the bfd sections as well, so we don't stop
|
||
with one match. */
|
||
bsec = abfd->sections;
|
||
while (bsec)
|
||
{
|
||
if (phdr[i].p_vaddr <= bsec->vma
|
||
&& bsec->vma <= phdr[i].p_vaddr + (phdr[i].p_filesz - 1))
|
||
{
|
||
bsec->lma = phdr[i].p_paddr + (bsec->vma - phdr[i].p_vaddr);
|
||
}
|
||
bsec = bsec->next;
|
||
}
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
|
||
#ifdef DEBUG
|
||
void
|
||
rx_dump_symtab (bfd * abfd, void * internal_syms, void * external_syms)
|
||
{
|
||
size_t locsymcount;
|
||
Elf_Internal_Sym * isymbuf;
|
||
Elf_Internal_Sym * isymend;
|
||
Elf_Internal_Sym * isym;
|
||
Elf_Internal_Shdr * symtab_hdr;
|
||
bfd_boolean free_internal = FALSE, free_external = FALSE;
|
||
char * st_info_str;
|
||
char * st_info_stb_str;
|
||
char * st_other_str;
|
||
char * st_shndx_str;
|
||
|
||
if (! internal_syms)
|
||
{
|
||
internal_syms = bfd_malloc (1000);
|
||
free_internal = 1;
|
||
}
|
||
if (! external_syms)
|
||
{
|
||
external_syms = bfd_malloc (1000);
|
||
free_external = 1;
|
||
}
|
||
|
||
symtab_hdr = &elf_tdata (abfd)->symtab_hdr;
|
||
locsymcount = symtab_hdr->sh_size / get_elf_backend_data (abfd)->s->sizeof_sym;
|
||
if (free_internal)
|
||
isymbuf = bfd_elf_get_elf_syms (abfd, symtab_hdr,
|
||
symtab_hdr->sh_info, 0,
|
||
internal_syms, external_syms, NULL);
|
||
else
|
||
isymbuf = internal_syms;
|
||
isymend = isymbuf + locsymcount;
|
||
|
||
for (isym = isymbuf ; isym < isymend ; isym++)
|
||
{
|
||
switch (ELF_ST_TYPE (isym->st_info))
|
||
{
|
||
case STT_FUNC: st_info_str = "STT_FUNC";
|
||
case STT_SECTION: st_info_str = "STT_SECTION";
|
||
case STT_FILE: st_info_str = "STT_FILE";
|
||
case STT_OBJECT: st_info_str = "STT_OBJECT";
|
||
case STT_TLS: st_info_str = "STT_TLS";
|
||
default: st_info_str = "";
|
||
}
|
||
switch (ELF_ST_BIND (isym->st_info))
|
||
{
|
||
case STB_LOCAL: st_info_stb_str = "STB_LOCAL";
|
||
case STB_GLOBAL: st_info_stb_str = "STB_GLOBAL";
|
||
default: st_info_stb_str = "";
|
||
}
|
||
switch (ELF_ST_VISIBILITY (isym->st_other))
|
||
{
|
||
case STV_DEFAULT: st_other_str = "STV_DEFAULT";
|
||
case STV_INTERNAL: st_other_str = "STV_INTERNAL";
|
||
case STV_PROTECTED: st_other_str = "STV_PROTECTED";
|
||
default: st_other_str = "";
|
||
}
|
||
switch (isym->st_shndx)
|
||
{
|
||
case SHN_ABS: st_shndx_str = "SHN_ABS";
|
||
case SHN_COMMON: st_shndx_str = "SHN_COMMON";
|
||
case SHN_UNDEF: st_shndx_str = "SHN_UNDEF";
|
||
default: st_shndx_str = "";
|
||
}
|
||
|
||
printf ("isym = %p st_value = %lx st_size = %lx st_name = (%lu) %s "
|
||
"st_info = (%d) %s %s st_other = (%d) %s st_shndx = (%d) %s\n",
|
||
isym,
|
||
(unsigned long) isym->st_value,
|
||
(unsigned long) isym->st_size,
|
||
isym->st_name,
|
||
bfd_elf_string_from_elf_section (abfd, symtab_hdr->sh_link,
|
||
isym->st_name),
|
||
isym->st_info, st_info_str, st_info_stb_str,
|
||
isym->st_other, st_other_str,
|
||
isym->st_shndx, st_shndx_str);
|
||
}
|
||
if (free_internal)
|
||
free (internal_syms);
|
||
if (free_external)
|
||
free (external_syms);
|
||
}
|
||
|
||
char *
|
||
rx_get_reloc (long reloc)
|
||
{
|
||
if (0 <= reloc && reloc < R_RX_max)
|
||
return rx_elf_howto_table[reloc].name;
|
||
return "";
|
||
}
|
||
#endif /* DEBUG */
|
||
|
||
|
||
/* We must take care to keep the on-disk copy of any code sections
|
||
that are fully linked swapped if the target is big endian, to match
|
||
the Renesas tools. */
|
||
|
||
/* The rule is: big endian object that are final-link executables,
|
||
have code sections stored with 32-bit words swapped relative to
|
||
what you'd get by default. */
|
||
|
||
static bfd_boolean
|
||
rx_get_section_contents (bfd * abfd,
|
||
sec_ptr section,
|
||
void * location,
|
||
file_ptr offset,
|
||
bfd_size_type count)
|
||
{
|
||
int exec = (abfd->flags & EXEC_P) ? 1 : 0;
|
||
int s_code = (section->flags & SEC_CODE) ? 1 : 0;
|
||
bfd_boolean rv;
|
||
|
||
#ifdef DJDEBUG
|
||
fprintf (stderr, "dj: get %ld %ld from %s %s e%d sc%d %08lx:%08lx\n",
|
||
(long) offset, (long) count, section->name,
|
||
bfd_big_endian(abfd) ? "be" : "le",
|
||
exec, s_code, (long unsigned) section->filepos,
|
||
(long unsigned) offset);
|
||
#endif
|
||
|
||
if (exec && s_code && bfd_big_endian (abfd))
|
||
{
|
||
char * cloc = (char *) location;
|
||
bfd_size_type cnt, end_cnt;
|
||
|
||
rv = TRUE;
|
||
|
||
/* Fetch and swap unaligned bytes at the beginning. */
|
||
if (offset % 4)
|
||
{
|
||
char buf[4];
|
||
|
||
rv = _bfd_generic_get_section_contents (abfd, section, buf,
|
||
(offset & -4), 4);
|
||
if (!rv)
|
||
return FALSE;
|
||
|
||
bfd_putb32 (bfd_getl32 (buf), buf);
|
||
|
||
cnt = 4 - (offset % 4);
|
||
if (cnt > count)
|
||
cnt = count;
|
||
|
||
memcpy (location, buf + (offset % 4), cnt);
|
||
|
||
count -= cnt;
|
||
offset += cnt;
|
||
cloc += count;
|
||
}
|
||
|
||
end_cnt = count % 4;
|
||
|
||
/* Fetch and swap the middle bytes. */
|
||
if (count >= 4)
|
||
{
|
||
rv = _bfd_generic_get_section_contents (abfd, section, cloc, offset,
|
||
count - end_cnt);
|
||
if (!rv)
|
||
return FALSE;
|
||
|
||
for (cnt = count; cnt >= 4; cnt -= 4, cloc += 4)
|
||
bfd_putb32 (bfd_getl32 (cloc), cloc);
|
||
}
|
||
|
||
/* Fetch and swap the end bytes. */
|
||
if (end_cnt > 0)
|
||
{
|
||
char buf[4];
|
||
|
||
/* Fetch the end bytes. */
|
||
rv = _bfd_generic_get_section_contents (abfd, section, buf,
|
||
offset + count - end_cnt, 4);
|
||
if (!rv)
|
||
return FALSE;
|
||
|
||
bfd_putb32 (bfd_getl32 (buf), buf);
|
||
memcpy (cloc, buf, end_cnt);
|
||
}
|
||
}
|
||
else
|
||
rv = _bfd_generic_get_section_contents (abfd, section, location, offset, count);
|
||
|
||
return rv;
|
||
}
|
||
|
||
#ifdef DJDEBUG
|
||
static bfd_boolean
|
||
rx2_set_section_contents (bfd * abfd,
|
||
sec_ptr section,
|
||
const void * location,
|
||
file_ptr offset,
|
||
bfd_size_type count)
|
||
{
|
||
bfd_size_type i;
|
||
|
||
fprintf (stderr, " set sec %s %08x loc %p offset %#x count %#x\n",
|
||
section->name, (unsigned) section->vma, location, (int) offset, (int) count);
|
||
for (i = 0; i < count; i++)
|
||
{
|
||
if (i % 16 == 0 && i > 0)
|
||
fprintf (stderr, "\n");
|
||
|
||
if (i % 16 && i % 4 == 0)
|
||
fprintf (stderr, " ");
|
||
|
||
if (i % 16 == 0)
|
||
fprintf (stderr, " %08x:", (int) (section->vma + offset + i));
|
||
|
||
fprintf (stderr, " %02x", ((unsigned char *) location)[i]);
|
||
}
|
||
fprintf (stderr, "\n");
|
||
|
||
return _bfd_elf_set_section_contents (abfd, section, location, offset, count);
|
||
}
|
||
#define _bfd_elf_set_section_contents rx2_set_section_contents
|
||
#endif
|
||
|
||
static bfd_boolean
|
||
rx_set_section_contents (bfd * abfd,
|
||
sec_ptr section,
|
||
const void * location,
|
||
file_ptr offset,
|
||
bfd_size_type count)
|
||
{
|
||
bfd_boolean exec = (abfd->flags & EXEC_P) ? TRUE : FALSE;
|
||
bfd_boolean s_code = (section->flags & SEC_CODE) ? TRUE : FALSE;
|
||
bfd_boolean rv;
|
||
char * swapped_data = NULL;
|
||
bfd_size_type i;
|
||
bfd_vma caddr = section->vma + offset;
|
||
file_ptr faddr = 0;
|
||
bfd_size_type scount;
|
||
|
||
#ifdef DJDEBUG
|
||
bfd_size_type i;
|
||
|
||
fprintf (stderr, "\ndj: set %ld %ld to %s %s e%d sc%d\n",
|
||
(long) offset, (long) count, section->name,
|
||
bfd_big_endian (abfd) ? "be" : "le",
|
||
exec, s_code);
|
||
|
||
for (i = 0; i < count; i++)
|
||
{
|
||
int a = section->vma + offset + i;
|
||
|
||
if (a % 16 == 0 && a > 0)
|
||
fprintf (stderr, "\n");
|
||
|
||
if (a % 16 && a % 4 == 0)
|
||
fprintf (stderr, " ");
|
||
|
||
if (a % 16 == 0 || i == 0)
|
||
fprintf (stderr, " %08x:", (int) (section->vma + offset + i));
|
||
|
||
fprintf (stderr, " %02x", ((unsigned char *) location)[i]);
|
||
}
|
||
|
||
fprintf (stderr, "\n");
|
||
#endif
|
||
|
||
if (! exec || ! s_code || ! bfd_big_endian (abfd))
|
||
return _bfd_elf_set_section_contents (abfd, section, location, offset, count);
|
||
|
||
while (count > 0 && caddr > 0 && caddr % 4)
|
||
{
|
||
switch (caddr % 4)
|
||
{
|
||
case 0: faddr = offset + 3; break;
|
||
case 1: faddr = offset + 1; break;
|
||
case 2: faddr = offset - 1; break;
|
||
case 3: faddr = offset - 3; break;
|
||
}
|
||
|
||
rv = _bfd_elf_set_section_contents (abfd, section, location, faddr, 1);
|
||
if (! rv)
|
||
return rv;
|
||
|
||
location ++;
|
||
offset ++;
|
||
count --;
|
||
caddr ++;
|
||
}
|
||
|
||
scount = (int)(count / 4) * 4;
|
||
if (scount > 0)
|
||
{
|
||
char * cloc = (char *) location;
|
||
|
||
swapped_data = (char *) bfd_alloc (abfd, count);
|
||
|
||
for (i = 0; i < count; i += 4)
|
||
{
|
||
bfd_vma v = bfd_getl32 (cloc + i);
|
||
bfd_putb32 (v, swapped_data + i);
|
||
}
|
||
|
||
rv = _bfd_elf_set_section_contents (abfd, section, swapped_data, offset, scount);
|
||
|
||
if (!rv)
|
||
return rv;
|
||
}
|
||
|
||
count -= scount;
|
||
location += scount;
|
||
offset += scount;
|
||
|
||
if (count > 0)
|
||
{
|
||
caddr = section->vma + offset;
|
||
while (count > 0)
|
||
{
|
||
switch (caddr % 4)
|
||
{
|
||
case 0: faddr = offset + 3; break;
|
||
case 1: faddr = offset + 1; break;
|
||
case 2: faddr = offset - 1; break;
|
||
case 3: faddr = offset - 3; break;
|
||
}
|
||
rv = _bfd_elf_set_section_contents (abfd, section, location, faddr, 1);
|
||
if (! rv)
|
||
return rv;
|
||
|
||
location ++;
|
||
offset ++;
|
||
count --;
|
||
caddr ++;
|
||
}
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
static bfd_boolean
|
||
rx_final_link (bfd * abfd, struct bfd_link_info * info)
|
||
{
|
||
asection * o;
|
||
|
||
for (o = abfd->sections; o != NULL; o = o->next)
|
||
{
|
||
#ifdef DJDEBUG
|
||
fprintf (stderr, "sec %s fl %x vma %lx lma %lx size %lx raw %lx\n",
|
||
o->name, o->flags, o->vma, o->lma, o->size, o->rawsize);
|
||
#endif
|
||
if (o->flags & SEC_CODE
|
||
&& bfd_big_endian (abfd)
|
||
&& o->size % 4)
|
||
{
|
||
#ifdef DJDEBUG
|
||
fprintf (stderr, "adjusting...\n");
|
||
#endif
|
||
o->size += 4 - (o->size % 4);
|
||
}
|
||
}
|
||
|
||
return bfd_elf_final_link (abfd, info);
|
||
}
|
||
|
||
static bfd_boolean
|
||
elf32_rx_modify_program_headers (bfd * abfd ATTRIBUTE_UNUSED,
|
||
struct bfd_link_info * info ATTRIBUTE_UNUSED)
|
||
{
|
||
const struct elf_backend_data * bed;
|
||
struct elf_obj_tdata * tdata;
|
||
Elf_Internal_Phdr * phdr;
|
||
unsigned int count;
|
||
unsigned int i;
|
||
|
||
bed = get_elf_backend_data (abfd);
|
||
tdata = elf_tdata (abfd);
|
||
phdr = tdata->phdr;
|
||
count = tdata->program_header_size / bed->s->sizeof_phdr;
|
||
|
||
if (ignore_lma)
|
||
for (i = count; i-- != 0;)
|
||
if (phdr[i].p_type == PT_LOAD)
|
||
{
|
||
/* The Renesas tools expect p_paddr to be zero. However,
|
||
there is no other way to store the writable data in ROM for
|
||
startup initialization. So, we let the linker *think*
|
||
we're using paddr and vaddr the "usual" way, but at the
|
||
last minute we move the paddr into the vaddr (which is what
|
||
the simulator uses) and zero out paddr. Note that this
|
||
does not affect the section headers, just the program
|
||
headers. We hope. */
|
||
phdr[i].p_vaddr = phdr[i].p_paddr;
|
||
#if 0 /* If we zero out p_paddr, then the LMA in the section table
|
||
becomes wrong. */
|
||
phdr[i].p_paddr = 0;
|
||
#endif
|
||
}
|
||
|
||
return TRUE;
|
||
}
|
||
|
||
#define ELF_ARCH bfd_arch_rx
|
||
#define ELF_MACHINE_CODE EM_RX
|
||
#define ELF_MAXPAGESIZE 0x1000
|
||
|
||
#define TARGET_BIG_SYM bfd_elf32_rx_be_vec
|
||
#define TARGET_BIG_NAME "elf32-rx-be"
|
||
|
||
#define TARGET_LITTLE_SYM bfd_elf32_rx_le_vec
|
||
#define TARGET_LITTLE_NAME "elf32-rx-le"
|
||
|
||
#define elf_info_to_howto_rel NULL
|
||
#define elf_info_to_howto rx_info_to_howto_rela
|
||
#define elf_backend_object_p rx_elf_object_p
|
||
#define elf_backend_relocate_section rx_elf_relocate_section
|
||
#define elf_symbol_leading_char ('_')
|
||
#define elf_backend_can_gc_sections 1
|
||
#define elf_backend_modify_program_headers elf32_rx_modify_program_headers
|
||
|
||
#define bfd_elf32_bfd_reloc_type_lookup rx_reloc_type_lookup
|
||
#define bfd_elf32_bfd_reloc_name_lookup rx_reloc_name_lookup
|
||
#define bfd_elf32_bfd_set_private_flags rx_elf_set_private_flags
|
||
#define bfd_elf32_bfd_merge_private_bfd_data rx_elf_merge_private_bfd_data
|
||
#define bfd_elf32_bfd_print_private_bfd_data rx_elf_print_private_bfd_data
|
||
#define bfd_elf32_get_section_contents rx_get_section_contents
|
||
#define bfd_elf32_set_section_contents rx_set_section_contents
|
||
#define bfd_elf32_bfd_final_link rx_final_link
|
||
#define bfd_elf32_bfd_relax_section elf32_rx_relax_section_wrapper
|
||
|
||
#include "elf32-target.h"
|
||
|
||
/* We define a second big-endian target that doesn't have the custom
|
||
section get/set hooks, for times when we want to preserve the
|
||
pre-swapped .text sections (like objcopy). */
|
||
|
||
#undef TARGET_BIG_SYM
|
||
#define TARGET_BIG_SYM bfd_elf32_rx_be_ns_vec
|
||
#undef TARGET_BIG_NAME
|
||
#define TARGET_BIG_NAME "elf32-rx-be-ns"
|
||
#undef TARGET_LITTLE_SYM
|
||
|
||
#undef bfd_elf32_get_section_contents
|
||
#undef bfd_elf32_set_section_contents
|
||
|
||
#undef elf32_bed
|
||
#define elf32_bed elf32_rx_be_ns_bed
|
||
|
||
#include "elf32-target.h"
|