Copyright 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007,
2008 Free Software Foundation, Inc.
This file is part of BFD, the Binary File Descriptor library.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 51 Franklin Street - Fifth Floor, Boston,
MA 02110-1301, USA. */
#include "sysdep.h"
#include "bfd.h"
#include "libiberty.h"
#include "libbfd.h"
#include "elf-bfd.h"
#include "elf-vxworks.h"
#include "elf/arm.h"
bfd's elf32_arm_link_hash_entry. */
#define RELOC_SECTION(HTAB, NAME) \
((HTAB)->use_rel ? ".rel" NAME : ".rela" NAME)
elf32_arm_link_hash_entry. */
#define RELOC_SIZE(HTAB) \
((HTAB)->use_rel \
? sizeof (Elf32_External_Rel) \
: sizeof (Elf32_External_Rela))
elf32_arm_link_hash_entry. */
#define SWAP_RELOC_IN(HTAB) \
((HTAB)->use_rel \
? bfd_elf32_swap_reloc_in \
: bfd_elf32_swap_reloca_in)
elf32_arm_link_hash_entry. */
#define SWAP_RELOC_OUT(HTAB) \
((HTAB)->use_rel \
? bfd_elf32_swap_reloc_out \
: bfd_elf32_swap_reloca_out)
#define elf_info_to_howto 0
#define elf_info_to_howto_rel elf32_arm_info_to_howto
#define ARM_ELF_ABI_VERSION 0
#define ARM_ELF_OS_ABI_VERSION ELFOSABI_ARM
static struct elf_backend_data elf32_arm_vxworks_bed;
R_ARM_PC24 as an index into this, and find the R_ARM_PC24 HOWTO
in that slot. */
static reloc_howto_type elf32_arm_howto_table_1[] =
{
HOWTO (R_ARM_NONE,
0,
0,
0,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_NONE",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_PC24,
2,
2,
24,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_PC24",
FALSE,
0x00ffffff,
0x00ffffff,
TRUE),
HOWTO (R_ARM_ABS32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_ABS32",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_REL32,
0,
2,
32,
TRUE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_REL32",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_PC_G0,
0,
0,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_PC_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ABS16,
0,
1,
16,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_ABS16",
FALSE,
0x0000ffff,
0x0000ffff,
FALSE),
HOWTO (R_ARM_ABS12,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_ABS12",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
HOWTO (R_ARM_THM_ABS5,
6,
1,
5,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_THM_ABS5",
FALSE,
0x000007e0,
0x000007e0,
FALSE),
HOWTO (R_ARM_ABS8,
0,
0,
8,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_ABS8",
FALSE,
0x000000ff,
0x000000ff,
FALSE),
HOWTO (R_ARM_SBREL32,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_SBREL32",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_THM_CALL,
1,
2,
25,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_CALL",
FALSE,
0x07ff07ff,
0x07ff07ff,
TRUE),
HOWTO (R_ARM_THM_PC8,
1,
1,
8,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_PC8",
FALSE,
0x000000ff,
0x000000ff,
TRUE),
HOWTO (R_ARM_BREL_ADJ,
1,
1,
32,
FALSE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_BREL_ADJ",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_SWI24,
0,
0,
0,
FALSE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_SWI24",
FALSE,
0x00000000,
0x00000000,
FALSE),
HOWTO (R_ARM_THM_SWI8,
0,
0,
0,
FALSE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_SWI8",
FALSE,
0x00000000,
0x00000000,
FALSE),
HOWTO (R_ARM_XPC25,
2,
2,
25,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_XPC25",
FALSE,
0x00ffffff,
0x00ffffff,
TRUE),
HOWTO (R_ARM_THM_XPC22,
2,
2,
22,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_XPC22",
FALSE,
0x07ff07ff,
0x07ff07ff,
TRUE),
HOWTO (R_ARM_TLS_DTPMOD32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_DTPMOD32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_DTPOFF32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_DTPOFF32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_TPOFF32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_TPOFF32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_COPY,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_COPY",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_GLOB_DAT,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GLOB_DAT",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_JUMP_SLOT,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_JUMP_SLOT",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_RELATIVE,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_RELATIVE",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_GOTOFF32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GOTOFF32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_GOTPC,
0,
2,
32,
TRUE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GOTPC",
TRUE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_GOT32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GOT32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_PLT32,
2,
2,
24,
TRUE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_PLT32",
FALSE,
0x00ffffff,
0x00ffffff,
TRUE),
HOWTO (R_ARM_CALL,
2,
2,
24,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_CALL",
FALSE,
0x00ffffff,
0x00ffffff,
TRUE),
HOWTO (R_ARM_JUMP24,
2,
2,
24,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_JUMP24",
FALSE,
0x00ffffff,
0x00ffffff,
TRUE),
HOWTO (R_ARM_THM_JUMP24,
1,
2,
24,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_JUMP24",
FALSE,
0x07ff2fff,
0x07ff2fff,
TRUE),
HOWTO (R_ARM_BASE_ABS,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_BASE_ABS",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_ALU_PCREL7_0,
0,
2,
12,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PCREL_7_0",
FALSE,
0x00000fff,
0x00000fff,
TRUE),
HOWTO (R_ARM_ALU_PCREL15_8,
0,
2,
12,
TRUE,
8,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PCREL_15_8",
FALSE,
0x00000fff,
0x00000fff,
TRUE),
HOWTO (R_ARM_ALU_PCREL23_15,
0,
2,
12,
TRUE,
16,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PCREL_23_15",
FALSE,
0x00000fff,
0x00000fff,
TRUE),
HOWTO (R_ARM_LDR_SBREL_11_0,
0,
2,
12,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_SBREL_11_0",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
HOWTO (R_ARM_ALU_SBREL_19_12,
0,
2,
8,
FALSE,
12,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SBREL_19_12",
FALSE,
0x000ff000,
0x000ff000,
FALSE),
HOWTO (R_ARM_ALU_SBREL_27_20,
0,
2,
8,
FALSE,
20,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SBREL_27_20",
FALSE,
0x0ff00000,
0x0ff00000,
FALSE),
HOWTO (R_ARM_TARGET1,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_TARGET1",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_ROSEGREL32,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ROSEGREL32",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_V4BX,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_V4BX",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TARGET2,
0,
2,
32,
FALSE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_TARGET2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_PREL31,
0,
2,
31,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_PREL31",
FALSE,
0x7fffffff,
0x7fffffff,
TRUE),
HOWTO (R_ARM_MOVW_ABS_NC,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_MOVW_ABS_NC",
FALSE,
0x000f0fff,
0x000f0fff,
FALSE),
HOWTO (R_ARM_MOVT_ABS,
0,
2,
16,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_MOVT_ABS",
FALSE,
0x000f0fff,
0x000f0fff,
FALSE),
HOWTO (R_ARM_MOVW_PREL_NC,
0,
2,
16,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_MOVW_PREL_NC",
FALSE,
0x000f0fff,
0x000f0fff,
TRUE),
HOWTO (R_ARM_MOVT_PREL,
0,
2,
16,
TRUE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_MOVT_PREL",
FALSE,
0x000f0fff,
0x000f0fff,
TRUE),
HOWTO (R_ARM_THM_MOVW_ABS_NC,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVW_ABS_NC",
FALSE,
0x040f70ff,
0x040f70ff,
FALSE),
HOWTO (R_ARM_THM_MOVT_ABS,
0,
2,
16,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVT_ABS",
FALSE,
0x040f70ff,
0x040f70ff,
FALSE),
HOWTO (R_ARM_THM_MOVW_PREL_NC,
0,
2,
16,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVW_PREL_NC",
FALSE,
0x040f70ff,
0x040f70ff,
TRUE),
HOWTO (R_ARM_THM_MOVT_PREL,
0,
2,
16,
TRUE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVT_PREL",
FALSE,
0x040f70ff,
0x040f70ff,
TRUE),
HOWTO (R_ARM_THM_JUMP19,
1,
2,
19,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_JUMP19",
FALSE,
0x043f2fff,
0x043f2fff,
TRUE),
HOWTO (R_ARM_THM_JUMP6,
1,
1,
6,
TRUE,
0,
complain_overflow_unsigned,
bfd_elf_generic_reloc,
"R_ARM_THM_JUMP6",
FALSE,
0x02f8,
0x02f8,
TRUE),
address -4095 .. 4095(base) by altering ADDW to SUBW or vice
versa. */
HOWTO (R_ARM_THM_ALU_PREL_11_0,
0,
2,
13,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_ALU_PREL_11_0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_THM_PC12,
0,
2,
13,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_PC12",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ABS32_NOI,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ABS32_NOI",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_REL32_NOI,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_REL32_NOI",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_ALU_PC_G0_NC,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PC_G0_NC",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_PC_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PC_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_PC_G1_NC,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PC_G1_NC",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_PC_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PC_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_PC_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_PC_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_PC_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_PC_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_PC_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_PC_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_PC_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_PC_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_PC_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_PC_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_PC_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_PC_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_PC_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_PC_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_PC_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_PC_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_PC_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_PC_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_SB_G0_NC,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SB_G0_NC",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_SB_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SB_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_SB_G1_NC,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SB_G1_NC",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_SB_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SB_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_ALU_SB_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_ALU_SB_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_SB_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_SB_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_SB_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_SB_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDR_SB_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDR_SB_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_SB_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_SB_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_SB_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_SB_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDRS_SB_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDRS_SB_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_SB_G0,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_SB_G0",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_SB_G1,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_SB_G1",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_LDC_SB_G2,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_LDC_SB_G2",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_MOVW_BREL_NC,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_MOVW_BREL_NC",
FALSE,
0x0000ffff,
0x0000ffff,
FALSE),
HOWTO (R_ARM_MOVT_BREL,
0,
2,
16,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_MOVT_BREL",
FALSE,
0x0000ffff,
0x0000ffff,
FALSE),
HOWTO (R_ARM_MOVW_BREL,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_MOVW_BREL",
FALSE,
0x0000ffff,
0x0000ffff,
FALSE),
HOWTO (R_ARM_THM_MOVW_BREL_NC,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVW_BREL_NC",
FALSE,
0x040f70ff,
0x040f70ff,
FALSE),
HOWTO (R_ARM_THM_MOVT_BREL,
0,
2,
16,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVT_BREL",
FALSE,
0x040f70ff,
0x040f70ff,
FALSE),
HOWTO (R_ARM_THM_MOVW_BREL,
0,
2,
16,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_THM_MOVW_BREL",
FALSE,
0x040f70ff,
0x040f70ff,
FALSE),
EMPTY_HOWTO (90),
EMPTY_HOWTO (91),
EMPTY_HOWTO (92),
EMPTY_HOWTO (93),
HOWTO (R_ARM_PLT32_ABS,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_PLT32_ABS",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_GOT_ABS,
0,
2,
32,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_GOT_ABS",
FALSE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_GOT_PREL,
0,
2,
32,
TRUE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_GOT_PREL",
FALSE,
0xffffffff,
0xffffffff,
TRUE),
HOWTO (R_ARM_GOT_BREL12,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GOT_BREL12",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
HOWTO (R_ARM_GOTOFF12,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_GOTOFF12",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
EMPTY_HOWTO (R_ARM_GOTRELAX),
HOWTO (R_ARM_GNU_VTENTRY,
0,
2,
0,
FALSE,
0,
complain_overflow_dont,
_bfd_elf_rel_vtable_reloc_fn,
"R_ARM_GNU_VTENTRY",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_GNU_VTINHERIT,
0,
2,
0,
FALSE,
0,
complain_overflow_dont,
NULL,
"R_ARM_GNU_VTINHERIT",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_THM_JUMP11,
1,
1,
11,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_JUMP11",
FALSE,
0x000007ff,
0x000007ff,
TRUE),
HOWTO (R_ARM_THM_JUMP8,
1,
1,
8,
TRUE,
0,
complain_overflow_signed,
bfd_elf_generic_reloc,
"R_ARM_THM_JUMP8",
FALSE,
0x000000ff,
0x000000ff,
TRUE),
HOWTO (R_ARM_TLS_GD32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
NULL,
"R_ARM_TLS_GD32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_LDM32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_LDM32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_LDO32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_LDO32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_IE32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
NULL,
"R_ARM_TLS_IE32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_LE32,
0,
2,
32,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_LE32",
TRUE,
0xffffffff,
0xffffffff,
FALSE),
HOWTO (R_ARM_TLS_LDO12,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_LDO12",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
HOWTO (R_ARM_TLS_LE12,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_LE12",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
HOWTO (R_ARM_TLS_IE12GP,
0,
2,
12,
FALSE,
0,
complain_overflow_bitfield,
bfd_elf_generic_reloc,
"R_ARM_TLS_IE12GP",
FALSE,
0x00000fff,
0x00000fff,
FALSE),
};
128 R_ARM_ME_TOO, obsolete
129-255 unallocated in AAELF.
249-255 extended, currently unused, relocations: */
static reloc_howto_type elf32_arm_howto_table_2[4] =
{
HOWTO (R_ARM_RREL32,
0,
0,
0,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_RREL32",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_RABS32,
0,
0,
0,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_RABS32",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_RPC24,
0,
0,
0,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_RPC24",
FALSE,
0,
0,
FALSE),
HOWTO (R_ARM_RBASE,
0,
0,
0,
FALSE,
0,
complain_overflow_dont,
bfd_elf_generic_reloc,
"R_ARM_RBASE",
FALSE,
0,
0,
FALSE)
};
static reloc_howto_type *
elf32_arm_howto_from_type (unsigned int r_type)
{
if (r_type < ARRAY_SIZE (elf32_arm_howto_table_1))
return &elf32_arm_howto_table_1[r_type];
if (r_type >= R_ARM_RREL32
&& r_type < R_ARM_RREL32 + ARRAY_SIZE (elf32_arm_howto_table_2))
return &elf32_arm_howto_table_2[r_type - R_ARM_RREL32];
return NULL;
}
static void
elf32_arm_info_to_howto (bfd * abfd ATTRIBUTE_UNUSED, arelent * bfd_reloc,
Elf_Internal_Rela * elf_reloc)
{
unsigned int r_type;
r_type = ELF32_R_TYPE (elf_reloc->r_info);
bfd_reloc->howto = elf32_arm_howto_from_type (r_type);
}
struct elf32_arm_reloc_map
{
bfd_reloc_code_real_type bfd_reloc_val;
unsigned char elf_reloc_val;
};
static const struct elf32_arm_reloc_map elf32_arm_reloc_map[] =
{
{BFD_RELOC_NONE, R_ARM_NONE},
{BFD_RELOC_ARM_PCREL_BRANCH, R_ARM_PC24},
{BFD_RELOC_ARM_PCREL_CALL, R_ARM_CALL},
{BFD_RELOC_ARM_PCREL_JUMP, R_ARM_JUMP24},
{BFD_RELOC_ARM_PCREL_BLX, R_ARM_XPC25},
{BFD_RELOC_THUMB_PCREL_BLX, R_ARM_THM_XPC22},
{BFD_RELOC_32, R_ARM_ABS32},
{BFD_RELOC_32_PCREL, R_ARM_REL32},
{BFD_RELOC_8, R_ARM_ABS8},
{BFD_RELOC_16, R_ARM_ABS16},
{BFD_RELOC_ARM_OFFSET_IMM, R_ARM_ABS12},
{BFD_RELOC_ARM_THUMB_OFFSET, R_ARM_THM_ABS5},
{BFD_RELOC_THUMB_PCREL_BRANCH25, R_ARM_THM_JUMP24},
{BFD_RELOC_THUMB_PCREL_BRANCH23, R_ARM_THM_CALL},
{BFD_RELOC_THUMB_PCREL_BRANCH12, R_ARM_THM_JUMP11},
{BFD_RELOC_THUMB_PCREL_BRANCH20, R_ARM_THM_JUMP19},
{BFD_RELOC_THUMB_PCREL_BRANCH9, R_ARM_THM_JUMP8},
{BFD_RELOC_THUMB_PCREL_BRANCH7, R_ARM_THM_JUMP6},
{BFD_RELOC_ARM_GLOB_DAT, R_ARM_GLOB_DAT},
{BFD_RELOC_ARM_JUMP_SLOT, R_ARM_JUMP_SLOT},
{BFD_RELOC_ARM_RELATIVE, R_ARM_RELATIVE},
{BFD_RELOC_ARM_GOTOFF, R_ARM_GOTOFF32},
{BFD_RELOC_ARM_GOTPC, R_ARM_GOTPC},
{BFD_RELOC_ARM_GOT32, R_ARM_GOT32},
{BFD_RELOC_ARM_PLT32, R_ARM_PLT32},
{BFD_RELOC_ARM_TARGET1, R_ARM_TARGET1},
{BFD_RELOC_ARM_ROSEGREL32, R_ARM_ROSEGREL32},
{BFD_RELOC_ARM_SBREL32, R_ARM_SBREL32},
{BFD_RELOC_ARM_PREL31, R_ARM_PREL31},
{BFD_RELOC_ARM_TARGET2, R_ARM_TARGET2},
{BFD_RELOC_ARM_PLT32, R_ARM_PLT32},
{BFD_RELOC_ARM_TLS_GD32, R_ARM_TLS_GD32},
{BFD_RELOC_ARM_TLS_LDO32, R_ARM_TLS_LDO32},
{BFD_RELOC_ARM_TLS_LDM32, R_ARM_TLS_LDM32},
{BFD_RELOC_ARM_TLS_DTPMOD32, R_ARM_TLS_DTPMOD32},
{BFD_RELOC_ARM_TLS_DTPOFF32, R_ARM_TLS_DTPOFF32},
{BFD_RELOC_ARM_TLS_TPOFF32, R_ARM_TLS_TPOFF32},
{BFD_RELOC_ARM_TLS_IE32, R_ARM_TLS_IE32},
{BFD_RELOC_ARM_TLS_LE32, R_ARM_TLS_LE32},
{BFD_RELOC_VTABLE_INHERIT, R_ARM_GNU_VTINHERIT},
{BFD_RELOC_VTABLE_ENTRY, R_ARM_GNU_VTENTRY},
{BFD_RELOC_ARM_MOVW, R_ARM_MOVW_ABS_NC},
{BFD_RELOC_ARM_MOVT, R_ARM_MOVT_ABS},
{BFD_RELOC_ARM_MOVW_PCREL, R_ARM_MOVW_PREL_NC},
{BFD_RELOC_ARM_MOVT_PCREL, R_ARM_MOVT_PREL},
{BFD_RELOC_ARM_THUMB_MOVW, R_ARM_THM_MOVW_ABS_NC},
{BFD_RELOC_ARM_THUMB_MOVT, R_ARM_THM_MOVT_ABS},
{BFD_RELOC_ARM_THUMB_MOVW_PCREL, R_ARM_THM_MOVW_PREL_NC},
{BFD_RELOC_ARM_THUMB_MOVT_PCREL, R_ARM_THM_MOVT_PREL},
{BFD_RELOC_ARM_ALU_PC_G0_NC, R_ARM_ALU_PC_G0_NC},
{BFD_RELOC_ARM_ALU_PC_G0, R_ARM_ALU_PC_G0},
{BFD_RELOC_ARM_ALU_PC_G1_NC, R_ARM_ALU_PC_G1_NC},
{BFD_RELOC_ARM_ALU_PC_G1, R_ARM_ALU_PC_G1},
{BFD_RELOC_ARM_ALU_PC_G2, R_ARM_ALU_PC_G2},
{BFD_RELOC_ARM_LDR_PC_G0, R_ARM_LDR_PC_G0},
{BFD_RELOC_ARM_LDR_PC_G1, R_ARM_LDR_PC_G1},
{BFD_RELOC_ARM_LDR_PC_G2, R_ARM_LDR_PC_G2},
{BFD_RELOC_ARM_LDRS_PC_G0, R_ARM_LDRS_PC_G0},
{BFD_RELOC_ARM_LDRS_PC_G1, R_ARM_LDRS_PC_G1},
{BFD_RELOC_ARM_LDRS_PC_G2, R_ARM_LDRS_PC_G2},
{BFD_RELOC_ARM_LDC_PC_G0, R_ARM_LDC_PC_G0},
{BFD_RELOC_ARM_LDC_PC_G1, R_ARM_LDC_PC_G1},
{BFD_RELOC_ARM_LDC_PC_G2, R_ARM_LDC_PC_G2},
{BFD_RELOC_ARM_ALU_SB_G0_NC, R_ARM_ALU_SB_G0_NC},
{BFD_RELOC_ARM_ALU_SB_G0, R_ARM_ALU_SB_G0},
{BFD_RELOC_ARM_ALU_SB_G1_NC, R_ARM_ALU_SB_G1_NC},
{BFD_RELOC_ARM_ALU_SB_G1, R_ARM_ALU_SB_G1},
{BFD_RELOC_ARM_ALU_SB_G2, R_ARM_ALU_SB_G2},
{BFD_RELOC_ARM_LDR_SB_G0, R_ARM_LDR_SB_G0},
{BFD_RELOC_ARM_LDR_SB_G1, R_ARM_LDR_SB_G1},
{BFD_RELOC_ARM_LDR_SB_G2, R_ARM_LDR_SB_G2},
{BFD_RELOC_ARM_LDRS_SB_G0, R_ARM_LDRS_SB_G0},
{BFD_RELOC_ARM_LDRS_SB_G1, R_ARM_LDRS_SB_G1},
{BFD_RELOC_ARM_LDRS_SB_G2, R_ARM_LDRS_SB_G2},
{BFD_RELOC_ARM_LDC_SB_G0, R_ARM_LDC_SB_G0},
{BFD_RELOC_ARM_LDC_SB_G1, R_ARM_LDC_SB_G1},
{BFD_RELOC_ARM_LDC_SB_G2, R_ARM_LDC_SB_G2},
{BFD_RELOC_ARM_V4BX, R_ARM_V4BX}
};
static reloc_howto_type *
elf32_arm_reloc_type_lookup (bfd *abfd ATTRIBUTE_UNUSED,
bfd_reloc_code_real_type code)
{
unsigned int i;
for (i = 0; i < ARRAY_SIZE (elf32_arm_reloc_map); i ++)
if (elf32_arm_reloc_map[i].bfd_reloc_val == code)
return elf32_arm_howto_from_type (elf32_arm_reloc_map[i].elf_reloc_val);
return NULL;
}
static reloc_howto_type *
elf32_arm_reloc_name_lookup (bfd *abfd ATTRIBUTE_UNUSED,
const char *r_name)
{
unsigned int i;
for (i = 0; i < ARRAY_SIZE (elf32_arm_howto_table_1); i++)
if (elf32_arm_howto_table_1[i].name != NULL
&& strcasecmp (elf32_arm_howto_table_1[i].name, r_name) == 0)
return &elf32_arm_howto_table_1[i];
for (i = 0; i < ARRAY_SIZE (elf32_arm_howto_table_2); i++)
if (elf32_arm_howto_table_2[i].name != NULL
&& strcasecmp (elf32_arm_howto_table_2[i].name, r_name) == 0)
return &elf32_arm_howto_table_2[i];
return NULL;
}
static bfd_boolean
elf32_arm_nabi_grok_prstatus (bfd *abfd, Elf_Internal_Note *note)
{
int offset;
size_t size;
switch (note->descsz)
{
default:
return FALSE;
case 148:
elf_tdata (abfd)->core_signal = bfd_get_16 (abfd, note->descdata + 12);
elf_tdata (abfd)->core_pid = bfd_get_32 (abfd, note->descdata + 24);
offset = 72;
size = 72;
break;
}
return _bfd_elfcore_make_pseudosection (abfd, ".reg",
size, note->descpos + offset);
}
static bfd_boolean
elf32_arm_nabi_grok_psinfo (bfd *abfd, Elf_Internal_Note *note)
{
switch (note->descsz)
{
default:
return FALSE;
case 124:
elf_tdata (abfd)->core_program
= _bfd_elfcore_strndup (abfd, note->descdata + 28, 16);
elf_tdata (abfd)->core_command
= _bfd_elfcore_strndup (abfd, note->descdata + 44, 80);
}
onto the end of the args in some (at least one anyway)
implementations, so strip it off if it exists. */
{
char *command = elf_tdata (abfd)->core_command;
int n = strlen (command);
if (0 < n && command[n - 1] == ' ')
command[n - 1] = '\0';
}
return TRUE;
}
#define TARGET_LITTLE_SYM bfd_elf32_littlearm_vec
#define TARGET_LITTLE_NAME "elf32-littlearm"
#define TARGET_BIG_SYM bfd_elf32_bigarm_vec
#define TARGET_BIG_NAME "elf32-bigarm"
#define elf_backend_grok_prstatus elf32_arm_nabi_grok_prstatus
#define elf_backend_grok_psinfo elf32_arm_nabi_grok_psinfo
typedef unsigned long int insn32;
typedef unsigned short int insn16;
interworkable. */
#define INTERWORK_FLAG(abfd) \
(EF_ARM_EABI_VERSION (elf_elfheader (abfd)->e_flags) >= EF_ARM_EABI_VER4 \
|| (elf_elfheader (abfd)->e_flags & EF_ARM_INTERWORK))
The entry_names are used to do simple name mangling on the stubs.
Given a function name, and its type, the stub can be found. The
name can be changed. The only requirement is the %s be present. */
#define THUMB2ARM_GLUE_SECTION_NAME ".glue_7t"
#define THUMB2ARM_GLUE_ENTRY_NAME "__%s_from_thumb"
#define ARM2THUMB_GLUE_SECTION_NAME ".glue_7"
#define ARM2THUMB_GLUE_ENTRY_NAME "__%s_from_arm"
#define VFP11_ERRATUM_VENEER_SECTION_NAME ".vfp11_veneer"
#define VFP11_ERRATUM_VENEER_ENTRY_NAME "__vfp11_veneer_%x"
#define ARM_BX_GLUE_SECTION_NAME ".v4_bx"
#define ARM_BX_GLUE_ENTRY_NAME "__bx_r%d"
#define STUB_ENTRY_NAME "__%s_veneer"
section. */
#define ELF_DYNAMIC_INTERPRETER "/usr/lib/ld.so.1"
#ifdef FOUR_WORD_PLT
this. It is set up so that any shared library function that is
called before the relocation has been set up calls the dynamic
linker first. */
static const bfd_vma elf32_arm_plt0_entry [] =
{
0xe52de004,
0xe59fe010,
0xe08fe00e,
0xe5bef008,
};
this. */
static const bfd_vma elf32_arm_plt_entry [] =
{
0xe28fc600,
0xe28cca00,
0xe5bcf000,
0x00000000,
};
#else
this. It is set up so that any shared library function that is
called before the relocation has been set up calls the dynamic
linker first. */
static const bfd_vma elf32_arm_plt0_entry [] =
{
0xe52de004,
0xe59fe004,
0xe08fe00e,
0xe5bef008,
0x00000000,
};
this. */
static const bfd_vma elf32_arm_plt_entry [] =
{
0xe28fc600,
0xe28cca00,
0xe5bcf000,
};
#endif
for a VxWorks executable. */
static const bfd_vma elf32_arm_vxworks_exec_plt0_entry[] =
{
0xe52dc008,
0xe59fc000,
0xe59cf008,
0x00000000,
};
static const bfd_vma elf32_arm_vxworks_exec_plt_entry[] =
{
0xe59fc000,
0xe59cf000,
0x00000000,
0xe59fc000,
0xea000000,
0x00000000,
};
static const bfd_vma elf32_arm_vxworks_shared_plt_entry[] =
{
0xe59fc000,
0xe79cf009,
0x00000000,
0xe59fc000,
0xe599f008,
0x00000000,
};
#define PLT_THUMB_STUB_SIZE 4
static const bfd_vma elf32_arm_plt_thumb_stub [] =
{
0x4778,
0x46c0
};
address spaces. */
static const bfd_vma elf32_arm_symbian_plt_entry [] =
{
0xe51ff004,
0x00000000,
};
#define ARM_MAX_FWD_BRANCH_OFFSET ((((1 << 23) - 1) << 2) + 8)
#define ARM_MAX_BWD_BRANCH_OFFSET ((-((1 << 23) << 2)) + 8)
#define THM_MAX_FWD_BRANCH_OFFSET ((1 << 22) -2 + 4)
#define THM_MAX_BWD_BRANCH_OFFSET (-(1 << 22) + 4)
#define THM2_MAX_FWD_BRANCH_OFFSET (((1 << 24) - 2) + 4)
#define THM2_MAX_BWD_BRANCH_OFFSET (-(1 << 24) + 4)
static const bfd_vma arm_long_branch_stub[] =
{
0xe51ff004,
0x00000000,
};
static const bfd_vma arm_thumb_v4t_long_branch_stub[] =
{
0xe59fc000,
0xe12fff1c,
0x00000000,
};
static const bfd_vma arm_thumb_thumb_long_branch_stub[] =
{
0x4e02b540,
0x473046fe,
0xbf00bd40,
0x00000000,
};
static const bfd_vma arm_thumb_arm_v4t_long_branch_stub[] =
{
0x4e03b540,
0x473046fe,
0xe8bd4040,
0xe12fff1e,
0x00000000,
};
static const bfd_vma arm_thumb_arm_v4t_short_branch_stub[] =
{
0x46c04778,
0xea000000,
};
static const bfd_vma arm_pic_long_branch_stub[] =
{
0xe59fc000,
0xe08ff00c,
0x00000000,
};
string. */
#define STUB_SUFFIX ".stub"
enum elf32_arm_stub_type
{
arm_stub_none,
arm_stub_long_branch,
arm_thumb_v4t_stub_long_branch,
arm_thumb_thumb_stub_long_branch,
arm_thumb_arm_v4t_stub_long_branch,
arm_thumb_arm_v4t_stub_short_branch,
arm_stub_pic_long_branch,
};
struct elf32_arm_stub_hash_entry
{
struct bfd_hash_entry root;
asection *stub_sec;
bfd_vma stub_offset;
value when building the stubs (so the stub knows where to jump). */
bfd_vma target_value;
asection *target_section;
enum elf32_arm_stub_type stub_type;
struct elf32_arm_link_hash_entry *h;
unsigned char st_type;
stub sections, the first input section in the group. */
asection *id_sec;
stub name in the hash table has to be unique; this does not, so
it can be friendlier. */
char *output_name;
};
code/data. */
typedef struct elf32_elf_section_map
{
bfd_vma vma;
char type;
}
elf32_arm_section_map;
typedef enum
{
VFP11_ERRATUM_BRANCH_TO_ARM_VENEER,
VFP11_ERRATUM_BRANCH_TO_THUMB_VENEER,
VFP11_ERRATUM_ARM_VENEER,
VFP11_ERRATUM_THUMB_VENEER
}
elf32_vfp11_erratum_type;
typedef struct elf32_vfp11_erratum_list
{
struct elf32_vfp11_erratum_list *next;
bfd_vma vma;
union
{
struct
{
struct elf32_vfp11_erratum_list *veneer;
unsigned int vfp_insn;
} b;
struct
{
struct elf32_vfp11_erratum_list *branch;
unsigned int id;
} v;
} u;
elf32_vfp11_erratum_type type;
}
elf32_vfp11_erratum_list;
typedef struct _arm_elf_section_data
{
struct bfd_elf_section_data elf;
unsigned int mapcount;
unsigned int mapsize;
elf32_arm_section_map *map;
unsigned int erratumcount;
elf32_vfp11_erratum_list *erratumlist;
}
_arm_elf_section_data;
#define elf32_arm_section_data(sec) \
((_arm_elf_section_data *) elf_section_data (sec))
#define TCB_SIZE 8
struct elf_arm_obj_tdata
{
struct elf_obj_tdata root;
char *local_got_tls_type;
int no_enum_size_warning;
int no_wchar_size_warning;
};
#define elf_arm_tdata(bfd) \
((struct elf_arm_obj_tdata *) (bfd)->tdata.any)
#define elf32_arm_local_got_tls_type(bfd) \
(elf_arm_tdata (bfd)->local_got_tls_type)
#define is_arm_elf(bfd) \
(bfd_get_flavour (bfd) == bfd_target_elf_flavour \
&& elf_tdata (bfd) != NULL \
&& elf_object_id (bfd) == ARM_ELF_TDATA)
static bfd_boolean
elf32_arm_mkobject (bfd *abfd)
{
return bfd_elf_allocate_object (abfd, sizeof (struct elf_arm_obj_tdata),
ARM_ELF_TDATA);
}
decides to copy in check_relocs for each symbol. This is so that
it can discard PC relative relocs if it doesn't need them when
linking with -Bsymbolic. We store the information in a field
extending the regular ELF linker hash table. */
for a given symbol. */
struct elf32_arm_relocs_copied
{
struct elf32_arm_relocs_copied * next;
asection * section;
bfd_size_type count;
bfd_size_type pc_count;
};
#define elf32_arm_hash_entry(ent) ((struct elf32_arm_link_hash_entry *)(ent))
struct elf32_arm_link_hash_entry
{
struct elf_link_hash_entry root;
struct elf32_arm_relocs_copied * relocs_copied;
so that we can emit the Thumb trampoline only if needed. */
bfd_signed_vma plt_thumb_refcount;
conversion, so record them separately. */
bfd_signed_vma plt_maybe_thumb_refcount;
used, we need to record the index into .got.plt instead of
recomputing it from the PLT offset. */
bfd_signed_vma plt_got_offset;
#define GOT_UNKNOWN 0
#define GOT_NORMAL 1
#define GOT_TLS_GD 2
#define GOT_TLS_IE 4
unsigned char tls_type;
symbols with Arm stubs. */
struct elf_link_hash_entry *export_glue;
symbol. */
struct elf32_arm_stub_hash_entry *stub_cache;
};
#define elf32_arm_link_hash_traverse(table, func, info) \
(elf_link_hash_traverse \
(&(table)->root, \
(bfd_boolean (*) (struct elf_link_hash_entry *, void *)) (func), \
(info)))
#define elf32_arm_hash_table(info) \
((struct elf32_arm_link_hash_table *) ((info)->hash))
#define arm_stub_hash_lookup(table, string, create, copy) \
((struct elf32_arm_stub_hash_entry *) \
bfd_hash_lookup ((table), (string), (create), (copy)))
struct elf32_arm_link_hash_table
{
struct elf_link_hash_table root;
bfd_size_type thumb_glue_size;
bfd_size_type arm_glue_size;
bfd_size_type bx_glue_size;
veneer has been populated. */
bfd_vma bx_glue_offset[15];
veneers. */
bfd_size_type vfp11_erratum_glue_size;
bfd * bfd_of_glue_owner;
int byteswap_code;
Nonzero if R_ARM_TARGET1 means R_ARM_REL32. */
int target1_is_rel;
int target2_reloc;
1 = Convert BX to MOV PC.
2 = Generate v4 interworing stubs. */
int fix_v4bx;
int use_blx;
VFP11 denorm erratum. */
bfd_arm_vfp11_fix vfp11_fix;
int num_vfp11_fixes;
int pic_veneer;
bfd_size_type plt_header_size;
bfd_size_type plt_entry_size;
int vxworks_p;
int symbian_p;
int use_rel;
asection *sgot;
asection *sgotplt;
asection *srelgot;
asection *splt;
asection *srelplt;
asection *sdynbss;
asection *srelbss;
asection *srelplt2;
union
{
bfd_signed_vma refcount;
bfd_vma offset;
} tls_ldm_got;
struct sym_sec_cache sym_sec;
bfd * obfd;
struct bfd_hash_table stub_hash_table;
bfd *stub_bfd;
asection * (*add_stub_section) (const char *, asection *);
void (*layout_sections_again) (void);
information on stub grouping. */
struct map_stub
{
attached. */
asection *link_sec;
asection *stub_sec;
} *stub_group;
unsigned int bfd_count;
int top_index;
asection **input_list;
};
static struct bfd_hash_entry *
elf32_arm_link_hash_newfunc (struct bfd_hash_entry * entry,
struct bfd_hash_table * table,
const char * string)
{
struct elf32_arm_link_hash_entry * ret =
(struct elf32_arm_link_hash_entry *) entry;
subclass. */
if (ret == NULL)
ret = bfd_hash_allocate (table, sizeof (struct elf32_arm_link_hash_entry));
if (ret == NULL)
return (struct bfd_hash_entry *) ret;
ret = ((struct elf32_arm_link_hash_entry *)
_bfd_elf_link_hash_newfunc ((struct bfd_hash_entry *) ret,
table, string));
if (ret != NULL)
{
ret->relocs_copied = NULL;
ret->tls_type = GOT_UNKNOWN;
ret->plt_thumb_refcount = 0;
ret->plt_maybe_thumb_refcount = 0;
ret->plt_got_offset = -1;
ret->export_glue = NULL;
ret->stub_cache = NULL;
}
return (struct bfd_hash_entry *) ret;
}
static struct bfd_hash_entry *
stub_hash_newfunc (struct bfd_hash_entry *entry,
struct bfd_hash_table *table,
const char *string)
{
subclass. */
if (entry == NULL)
{
entry = bfd_hash_allocate (table,
sizeof (struct elf32_arm_stub_hash_entry));
if (entry == NULL)
return entry;
}
entry = bfd_hash_newfunc (entry, table, string);
if (entry != NULL)
{
struct elf32_arm_stub_hash_entry *eh;
eh = (struct elf32_arm_stub_hash_entry *) entry;
eh->stub_sec = NULL;
eh->stub_offset = 0;
eh->target_value = 0;
eh->target_section = NULL;
eh->stub_type = arm_stub_none;
eh->h = NULL;
eh->id_sec = NULL;
}
return entry;
}
with S. */
static bfd_boolean
reloc_section_p (struct elf32_arm_link_hash_table *htab,
const char *name, asection *s)
{
if (htab->use_rel)
return CONST_STRNEQ (name, ".rel") && strcmp (s->name, name + 4) == 0;
else
return CONST_STRNEQ (name, ".rela") && strcmp (s->name, name + 5) == 0;
}
shortcuts to them in our hash table. */
static bfd_boolean
create_got_section (bfd *dynobj, struct bfd_link_info *info)
{
struct elf32_arm_link_hash_table *htab;
htab = elf32_arm_hash_table (info);
if (htab->symbian_p)
return TRUE;
if (! _bfd_elf_create_got_section (dynobj, info))
return FALSE;
htab->sgot = bfd_get_section_by_name (dynobj, ".got");
htab->sgotplt = bfd_get_section_by_name (dynobj, ".got.plt");
if (!htab->sgot || !htab->sgotplt)
abort ();
htab->srelgot = bfd_make_section_with_flags (dynobj,
RELOC_SECTION (htab, ".got"),
(SEC_ALLOC | SEC_LOAD
| SEC_HAS_CONTENTS
| SEC_IN_MEMORY
| SEC_LINKER_CREATED
| SEC_READONLY));
if (htab->srelgot == NULL
|| ! bfd_set_section_alignment (dynobj, htab->srelgot, 2))
return FALSE;
return TRUE;
}
.rel(a).bss sections in DYNOBJ, and set up shortcuts to them in our
hash table. */
static bfd_boolean
elf32_arm_create_dynamic_sections (bfd *dynobj, struct bfd_link_info *info)
{
struct elf32_arm_link_hash_table *htab;
htab = elf32_arm_hash_table (info);
if (!htab->sgot && !create_got_section (dynobj, info))
return FALSE;
if (!_bfd_elf_create_dynamic_sections (dynobj, info))
return FALSE;
htab->splt = bfd_get_section_by_name (dynobj, ".plt");
htab->srelplt = bfd_get_section_by_name (dynobj,
RELOC_SECTION (htab, ".plt"));
htab->sdynbss = bfd_get_section_by_name (dynobj, ".dynbss");
if (!info->shared)
htab->srelbss = bfd_get_section_by_name (dynobj,
RELOC_SECTION (htab, ".bss"));
if (htab->vxworks_p)
{
if (!elf_vxworks_create_dynamic_sections (dynobj, info, &htab->srelplt2))
return FALSE;
if (info->shared)
{
htab->plt_header_size = 0;
htab->plt_entry_size
= 4 * ARRAY_SIZE (elf32_arm_vxworks_shared_plt_entry);
}
else
{
htab->plt_header_size
= 4 * ARRAY_SIZE (elf32_arm_vxworks_exec_plt0_entry);
htab->plt_entry_size
= 4 * ARRAY_SIZE (elf32_arm_vxworks_exec_plt_entry);
}
}
if (!htab->splt
|| !htab->srelplt
|| !htab->sdynbss
|| (!info->shared && !htab->srelbss))
abort ();
return TRUE;
}
static void
elf32_arm_copy_indirect_symbol (struct bfd_link_info *info,
struct elf_link_hash_entry *dir,
struct elf_link_hash_entry *ind)
{
struct elf32_arm_link_hash_entry *edir, *eind;
edir = (struct elf32_arm_link_hash_entry *) dir;
eind = (struct elf32_arm_link_hash_entry *) ind;
if (eind->relocs_copied != NULL)
{
if (edir->relocs_copied != NULL)
{
struct elf32_arm_relocs_copied **pp;
struct elf32_arm_relocs_copied *p;
list. Merge any entries against the same section. */
for (pp = &eind->relocs_copied; (p = *pp) != NULL; )
{
struct elf32_arm_relocs_copied *q;
for (q = edir->relocs_copied; q != NULL; q = q->next)
if (q->section == p->section)
{
q->pc_count += p->pc_count;
q->count += p->count;
*pp = p->next;
break;
}
if (q == NULL)
pp = &p->next;
}
*pp = edir->relocs_copied;
}
edir->relocs_copied = eind->relocs_copied;
eind->relocs_copied = NULL;
}
if (ind->root.type == bfd_link_hash_indirect)
{
edir->plt_thumb_refcount += eind->plt_thumb_refcount;
eind->plt_thumb_refcount = 0;
edir->plt_maybe_thumb_refcount += eind->plt_maybe_thumb_refcount;
eind->plt_maybe_thumb_refcount = 0;
if (dir->got.refcount <= 0)
{
edir->tls_type = eind->tls_type;
eind->tls_type = GOT_UNKNOWN;
}
}
_bfd_elf_link_hash_copy_indirect (info, dir, ind);
}
static struct bfd_link_hash_table *
elf32_arm_link_hash_table_create (bfd *abfd)
{
struct elf32_arm_link_hash_table *ret;
bfd_size_type amt = sizeof (struct elf32_arm_link_hash_table);
ret = bfd_malloc (amt);
if (ret == NULL)
return NULL;
if (!_bfd_elf_link_hash_table_init (& ret->root, abfd,
elf32_arm_link_hash_newfunc,
sizeof (struct elf32_arm_link_hash_entry)))
{
free (ret);
return NULL;
}
ret->sgot = NULL;
ret->sgotplt = NULL;
ret->srelgot = NULL;
ret->splt = NULL;
ret->srelplt = NULL;
ret->sdynbss = NULL;
ret->srelbss = NULL;
ret->srelplt2 = NULL;
ret->thumb_glue_size = 0;
ret->arm_glue_size = 0;
ret->bx_glue_size = 0;
memset (ret->bx_glue_offset, 0, sizeof (ret->bx_glue_offset));
ret->vfp11_fix = BFD_ARM_VFP11_FIX_NONE;
ret->vfp11_erratum_glue_size = 0;
ret->num_vfp11_fixes = 0;
ret->bfd_of_glue_owner = NULL;
ret->byteswap_code = 0;
ret->target1_is_rel = 0;
ret->target2_reloc = R_ARM_NONE;
#ifdef FOUR_WORD_PLT
ret->plt_header_size = 16;
ret->plt_entry_size = 16;
#else
ret->plt_header_size = 20;
ret->plt_entry_size = 12;
#endif
ret->fix_v4bx = 0;
ret->use_blx = 0;
ret->vxworks_p = 0;
ret->symbian_p = 0;
ret->use_rel = 1;
ret->sym_sec.abfd = NULL;
ret->obfd = abfd;
ret->tls_ldm_got.refcount = 0;
ret->stub_bfd = NULL;
ret->add_stub_section = NULL;
ret->layout_sections_again = NULL;
ret->stub_group = NULL;
ret->bfd_count = 0;
ret->top_index = 0;
ret->input_list = NULL;
if (!bfd_hash_table_init (&ret->stub_hash_table, stub_hash_newfunc,
sizeof (struct elf32_arm_stub_hash_entry)))
{
free (ret);
return NULL;
}
return &ret->root.root;
}
static void
elf32_arm_hash_table_free (struct bfd_link_hash_table *hash)
{
struct elf32_arm_link_hash_table *ret
= (struct elf32_arm_link_hash_table *) hash;
bfd_hash_table_free (&ret->stub_hash_table);
_bfd_generic_link_hash_table_free (hash);
}
static bfd_boolean
using_thumb_only (struct elf32_arm_link_hash_table *globals)
{
int arch = bfd_elf_get_obj_attr_int (globals->obfd, OBJ_ATTR_PROC,
Tag_CPU_arch);
int profile;
if (arch != TAG_CPU_ARCH_V7)
return FALSE;
profile = bfd_elf_get_obj_attr_int (globals->obfd, OBJ_ATTR_PROC,
Tag_CPU_arch_profile);
return profile == 'M';
}
static bfd_boolean
using_thumb2 (struct elf32_arm_link_hash_table *globals)
{
int arch = bfd_elf_get_obj_attr_int (globals->obfd, OBJ_ATTR_PROC,
Tag_CPU_arch);
return arch == TAG_CPU_ARCH_V6T2 || arch >= TAG_CPU_ARCH_V7;
}
static bfd_boolean
arm_stub_is_thumb (enum elf32_arm_stub_type stub_type)
{
switch (stub_type)
{
case arm_thumb_thumb_stub_long_branch:
case arm_thumb_arm_v4t_stub_long_branch:
case arm_thumb_arm_v4t_stub_short_branch:
return TRUE;
case arm_stub_none:
BFD_FAIL ();
return FALSE;
break;
default:
return FALSE;
}
}
static enum elf32_arm_stub_type
arm_type_of_stub (struct bfd_link_info *info,
asection *input_sec,
const Elf_Internal_Rela *rel,
unsigned char st_type,
struct elf32_arm_link_hash_entry *hash,
bfd_vma destination,
asection *sym_sec,
bfd *input_bfd,
const char *name)
{
bfd_vma location;
bfd_signed_vma branch_offset;
unsigned int r_type;
struct elf32_arm_link_hash_table * globals;
int thumb2;
int thumb_only;
enum elf32_arm_stub_type stub_type = arm_stub_none;
type STT_SECTION: give up. */
if (st_type == STT_SECTION)
return stub_type;
globals = elf32_arm_hash_table (info);
thumb_only = using_thumb_only (globals);
thumb2 = using_thumb2 (globals);
location = (input_sec->output_offset
+ input_sec->output_section->vma
+ rel->r_offset);
branch_offset = (bfd_signed_vma)(destination - location);
r_type = ELF32_R_TYPE (rel->r_info);
glue. */
if (globals->splt != NULL && hash != NULL && hash->root.plt.offset != (bfd_vma) -1)
return stub_type;
if (r_type == R_ARM_THM_CALL)
{
if ((!thumb2
&& (branch_offset > THM_MAX_FWD_BRANCH_OFFSET
|| (branch_offset < THM_MAX_BWD_BRANCH_OFFSET)))
|| (thumb2
&& (branch_offset > THM2_MAX_FWD_BRANCH_OFFSET
|| (branch_offset < THM2_MAX_BWD_BRANCH_OFFSET)))
|| ((st_type != STT_ARM_TFUNC) && !globals->use_blx))
{
if (st_type == STT_ARM_TFUNC)
{
if (!thumb_only)
{
stub_type = (info->shared | globals->pic_veneer)
? ((globals->use_blx)
? arm_stub_pic_long_branch
: arm_stub_none)
: (globals->use_blx)
? arm_stub_long_branch
: arm_stub_none;
}
else
{
stub_type = (info->shared | globals->pic_veneer)
? arm_stub_none
: (globals->use_blx)
? arm_thumb_thumb_stub_long_branch
: arm_stub_none;
}
}
else
{
if (sym_sec != NULL
&& sym_sec->owner != NULL
&& !INTERWORK_FLAG (sym_sec->owner))
{
(*_bfd_error_handler)
(_("%B(%s): warning: interworking not enabled.\n"
" first occurrence: %B: Thumb call to ARM"),
sym_sec->owner, input_bfd, name);
}
stub_type = (info->shared | globals->pic_veneer)
? ((globals->use_blx)
? arm_stub_pic_long_branch
: arm_stub_none)
: (globals->use_blx)
? arm_stub_long_branch
: arm_thumb_arm_v4t_stub_long_branch;
if ((stub_type == arm_thumb_arm_v4t_stub_long_branch)
&& (branch_offset <= THM_MAX_FWD_BRANCH_OFFSET)
&& (branch_offset >= THM_MAX_BWD_BRANCH_OFFSET))
stub_type = arm_thumb_arm_v4t_stub_short_branch;
}
}
}
else if (r_type == R_ARM_CALL)
{
if (st_type == STT_ARM_TFUNC)
{
if (sym_sec != NULL
&& sym_sec->owner != NULL
&& !INTERWORK_FLAG (sym_sec->owner))
{
(*_bfd_error_handler)
(_("%B(%s): warning: interworking not enabled.\n"
" first occurrence: %B: Thumb call to ARM"),
sym_sec->owner, input_bfd, name);
}
the mode change (bit 24 (H) of BLX encoding). */
if (branch_offset > (ARM_MAX_FWD_BRANCH_OFFSET + 2)
|| (branch_offset < ARM_MAX_BWD_BRANCH_OFFSET)
|| !globals->use_blx)
{
stub_type = (info->shared | globals->pic_veneer)
? arm_stub_pic_long_branch
: (globals->use_blx)
? arm_stub_long_branch
: arm_thumb_v4t_stub_long_branch;
}
}
else
{
if (branch_offset > ARM_MAX_FWD_BRANCH_OFFSET
|| (branch_offset < ARM_MAX_BWD_BRANCH_OFFSET))
{
stub_type = (info->shared | globals->pic_veneer)
? arm_stub_pic_long_branch
: arm_stub_long_branch;
}
}
}
return stub_type;
}
static char *
elf32_arm_stub_name (const asection *input_section,
const asection *sym_sec,
const struct elf32_arm_link_hash_entry *hash,
const Elf_Internal_Rela *rel)
{
char *stub_name;
bfd_size_type len;
if (hash)
{
len = 8 + 1 + strlen (hash->root.root.root.string) + 1 + 8 + 1;
stub_name = bfd_malloc (len);
if (stub_name != NULL)
sprintf (stub_name, "%08x_%s+%x",
input_section->id & 0xffffffff,
hash->root.root.root.string,
(int) rel->r_addend & 0xffffffff);
}
else
{
len = 8 + 1 + 8 + 1 + 8 + 1 + 8 + 1;
stub_name = bfd_malloc (len);
if (stub_name != NULL)
sprintf (stub_name, "%08x_%x:%x+%x",
input_section->id & 0xffffffff,
sym_sec->id & 0xffffffff,
(int) ELF32_R_SYM (rel->r_info) & 0xffffffff,
(int) rel->r_addend & 0xffffffff);
}
return stub_name;
}
creating the stub name takes a bit of time. */
static struct elf32_arm_stub_hash_entry *
elf32_arm_get_stub_entry (const asection *input_section,
const asection *sym_sec,
struct elf_link_hash_entry *hash,
const Elf_Internal_Rela *rel,
struct elf32_arm_link_hash_table *htab)
{
struct elf32_arm_stub_hash_entry *stub_entry;
struct elf32_arm_link_hash_entry *h = (struct elf32_arm_link_hash_entry *) hash;
const asection *id_sec;
if ((input_section->flags & SEC_CODE) == 0)
return NULL;
stub section, then use the id of the first section in the group.
Stub names need to include a section id, as there may well be
more than one stub used to reach say, printf, and we need to
distinguish between them. */
id_sec = htab->stub_group[input_section->id].link_sec;
if (h != NULL && h->stub_cache != NULL
&& h->stub_cache->h == h
&& h->stub_cache->id_sec == id_sec)
{
stub_entry = h->stub_cache;
}
else
{
char *stub_name;
stub_name = elf32_arm_stub_name (id_sec, sym_sec, h, rel);
if (stub_name == NULL)
return NULL;
stub_entry = arm_stub_hash_lookup (&htab->stub_hash_table,
stub_name, FALSE, FALSE);
if (h != NULL)
h->stub_cache = stub_entry;
free (stub_name);
}
return stub_entry;
}
stub entry are initialised. */
static struct elf32_arm_stub_hash_entry *
elf32_arm_add_stub (const char *stub_name,
asection *section,
struct elf32_arm_link_hash_table *htab)
{
asection *link_sec;
asection *stub_sec;
struct elf32_arm_stub_hash_entry *stub_entry;
link_sec = htab->stub_group[section->id].link_sec;
stub_sec = htab->stub_group[section->id].stub_sec;
if (stub_sec == NULL)
{
stub_sec = htab->stub_group[link_sec->id].stub_sec;
if (stub_sec == NULL)
{
size_t namelen;
bfd_size_type len;
char *s_name;
namelen = strlen (link_sec->name);
len = namelen + sizeof (STUB_SUFFIX);
s_name = bfd_alloc (htab->stub_bfd, len);
if (s_name == NULL)
return NULL;
memcpy (s_name, link_sec->name, namelen);
memcpy (s_name + namelen, STUB_SUFFIX, sizeof (STUB_SUFFIX));
stub_sec = (*htab->add_stub_section) (s_name, link_sec);
if (stub_sec == NULL)
return NULL;
htab->stub_group[link_sec->id].stub_sec = stub_sec;
}
htab->stub_group[section->id].stub_sec = stub_sec;
}
stub_entry = arm_stub_hash_lookup (&htab->stub_hash_table, stub_name,
TRUE, FALSE);
if (stub_entry == NULL)
{
(*_bfd_error_handler) (_("%s: cannot create stub entry %s"),
section->owner,
stub_name);
return NULL;
}
stub_entry->stub_sec = stub_sec;
stub_entry->stub_offset = 0;
stub_entry->id_sec = link_sec;
return stub_entry;
}
elf32_arm_write_section. */
static void
put_arm_insn (struct elf32_arm_link_hash_table * htab,
bfd * output_bfd, bfd_vma val, void * ptr)
{
if (htab->byteswap_code != bfd_little_endian (output_bfd))
bfd_putl32 (val, ptr);
else
bfd_putb32 (val, ptr);
}
elf32_arm_write_section. */
static void
put_thumb_insn (struct elf32_arm_link_hash_table * htab,
bfd * output_bfd, bfd_vma val, void * ptr)
{
if (htab->byteswap_code != bfd_little_endian (output_bfd))
bfd_putl16 (val, ptr);
else
bfd_putb16 (val, ptr);
}
static bfd_boolean
arm_build_one_stub (struct bfd_hash_entry *gen_entry,
void * in_arg)
{
struct elf32_arm_stub_hash_entry *stub_entry;
struct bfd_link_info *info;
struct elf32_arm_link_hash_table *htab;
asection *stub_sec;
bfd *stub_bfd;
bfd_vma stub_addr;
bfd_byte *loc;
bfd_vma sym_value;
int template_size;
int size;
const bfd_vma *template;
int i;
struct elf32_arm_link_hash_table * globals;
stub_entry = (struct elf32_arm_stub_hash_entry *) gen_entry;
info = (struct bfd_link_info *) in_arg;
globals = elf32_arm_hash_table (info);
htab = elf32_arm_hash_table (info);
stub_sec = stub_entry->stub_sec;
stub_entry->stub_offset = stub_sec->size;
loc = stub_sec->contents + stub_entry->stub_offset;
stub_bfd = stub_sec->owner;
stub_addr = stub_sec->output_section->vma + stub_sec->output_offset
+ stub_entry->stub_offset;
sym_value = (stub_entry->target_value
+ stub_entry->target_section->output_offset
+ stub_entry->target_section->output_section->vma);
switch (stub_entry->stub_type)
{
case arm_stub_long_branch:
template = arm_long_branch_stub;
template_size = (sizeof (arm_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_v4t_stub_long_branch:
template = arm_thumb_v4t_long_branch_stub;
template_size = (sizeof (arm_thumb_v4t_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_thumb_stub_long_branch:
template = arm_thumb_thumb_long_branch_stub;
template_size = (sizeof (arm_thumb_thumb_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_arm_v4t_stub_long_branch:
template = arm_thumb_arm_v4t_long_branch_stub;
template_size = (sizeof (arm_thumb_arm_v4t_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_arm_v4t_stub_short_branch:
template = arm_thumb_arm_v4t_short_branch_stub;
template_size = (sizeof(arm_thumb_arm_v4t_short_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_stub_pic_long_branch:
template = arm_pic_long_branch_stub;
template_size = (sizeof (arm_pic_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
default:
BFD_FAIL ();
return FALSE;
}
size = 0;
for (i = 0; i < (template_size / 4); i++)
{
instruction. */
if (template[i] != 0)
put_arm_insn (globals, stub_bfd, template[i], loc + size);
else
bfd_put_32 (stub_bfd, template[i], loc + size);
size += 4;
}
stub_sec->size += size;
if (stub_entry->st_type == STT_ARM_TFUNC)
sym_value |= 1;
switch (stub_entry->stub_type)
{
case arm_stub_long_branch:
_bfd_final_link_relocate (elf32_arm_howto_from_type (R_ARM_ABS32),
stub_bfd, stub_sec, stub_sec->contents,
stub_entry->stub_offset + 4, sym_value, 0);
break;
case arm_thumb_v4t_stub_long_branch:
_bfd_final_link_relocate (elf32_arm_howto_from_type (R_ARM_ABS32),
stub_bfd, stub_sec, stub_sec->contents,
stub_entry->stub_offset + 8, sym_value, 0);
break;
case arm_thumb_thumb_stub_long_branch:
_bfd_final_link_relocate (elf32_arm_howto_from_type (R_ARM_ABS32),
stub_bfd, stub_sec, stub_sec->contents,
stub_entry->stub_offset + 12, sym_value, 0);
break;
case arm_thumb_arm_v4t_stub_long_branch:
_bfd_final_link_relocate (elf32_arm_howto_from_type (R_ARM_ABS32),
stub_bfd, stub_sec, stub_sec->contents,
stub_entry->stub_offset + 16, sym_value, 0);
break;
case arm_thumb_arm_v4t_stub_short_branch:
{
long int rel_offset;
static const insn32 t2a3_b_insn = 0xea000000;
rel_offset = sym_value - (stub_addr + 8 + 4);
put_arm_insn (globals, stub_bfd,
(bfd_vma) t2a3_b_insn | ((rel_offset >> 2) & 0x00FFFFFF),
loc + 4);
}
break;
case arm_stub_pic_long_branch:
start of the stub. */
_bfd_final_link_relocate (elf32_arm_howto_from_type (R_ARM_REL32),
stub_bfd, stub_sec, stub_sec->contents,
stub_entry->stub_offset + 8, sym_value, 0);
break;
default:
break;
}
return TRUE;
}
we know stub section sizes. */
static bfd_boolean
arm_size_one_stub (struct bfd_hash_entry *gen_entry,
void * in_arg)
{
struct elf32_arm_stub_hash_entry *stub_entry;
struct elf32_arm_link_hash_table *htab;
const bfd_vma *template;
int template_size;
int size;
int i;
stub_entry = (struct elf32_arm_stub_hash_entry *) gen_entry;
htab = (struct elf32_arm_link_hash_table *) in_arg;
switch (stub_entry->stub_type)
{
case arm_stub_long_branch:
template = arm_long_branch_stub;
template_size = (sizeof (arm_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_v4t_stub_long_branch:
template = arm_thumb_v4t_long_branch_stub;
template_size = (sizeof (arm_thumb_v4t_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_thumb_stub_long_branch:
template = arm_thumb_thumb_long_branch_stub;
template_size = (sizeof (arm_thumb_thumb_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_arm_v4t_stub_long_branch:
template = arm_thumb_arm_v4t_long_branch_stub;
template_size = (sizeof (arm_thumb_arm_v4t_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_thumb_arm_v4t_stub_short_branch:
template = arm_thumb_arm_v4t_short_branch_stub;
template_size = (sizeof(arm_thumb_arm_v4t_short_branch_stub) / sizeof (bfd_vma)) * 4;
break;
case arm_stub_pic_long_branch:
template = arm_pic_long_branch_stub;
template_size = (sizeof (arm_pic_long_branch_stub) / sizeof (bfd_vma)) * 4;
break;
default:
BFD_FAIL ();
return FALSE;
break;
}
size = 0;
for (i = 0; i < (template_size / 4); i++)
size += 4;
size = (size + 7) & ~7;
stub_entry->stub_sec->size += size;
return TRUE;
}
for each output section included in the link. Returns -1 on error,
0 when no stubs will be needed, and 1 on success. */
int
elf32_arm_setup_section_lists (bfd *output_bfd,
struct bfd_link_info *info)
{
bfd *input_bfd;
unsigned int bfd_count;
int top_id, top_index;
asection *section;
asection **input_list, **list;
bfd_size_type amt;
struct elf32_arm_link_hash_table *htab = elf32_arm_hash_table (info);
if (! is_elf_hash_table (htab))
return 0;
for (input_bfd = info->input_bfds, bfd_count = 0, top_id = 0;
input_bfd != NULL;
input_bfd = input_bfd->link_next)
{
bfd_count += 1;
for (section = input_bfd->sections;
section != NULL;
section = section->next)
{
if (top_id < section->id)
top_id = section->id;
}
}
htab->bfd_count = bfd_count;
amt = sizeof (struct map_stub) * (top_id + 1);
htab->stub_group = bfd_zmalloc (amt);
if (htab->stub_group == NULL)
return -1;
section index as some sections may have been removed, and
_bfd_strip_section_from_output doesn't renumber the indices. */
for (section = output_bfd->sections, top_index = 0;
section != NULL;
section = section->next)
{
if (top_index < section->index)
top_index = section->index;
}
htab->top_index = top_index;
amt = sizeof (asection *) * (top_index + 1);
input_list = bfd_malloc (amt);
htab->input_list = input_list;
if (input_list == NULL)
return -1;
value we can check later. */
list = input_list + top_index;
do
*list = bfd_abs_section_ptr;
while (list-- != input_list);
for (section = output_bfd->sections;
section != NULL;
section = section->next)
{
if ((section->flags & SEC_CODE) != 0)
input_list[section->index] = NULL;
}
return 1;
}
in the order that input sections are linked into output sections.
Build lists of input sections to determine groupings between which
we may insert linker stubs. */
void
elf32_arm_next_input_section (struct bfd_link_info *info,
asection *isec)
{
struct elf32_arm_link_hash_table *htab = elf32_arm_hash_table (info);
if (isec->output_section->index <= htab->top_index)
{
asection **list = htab->input_list + isec->output_section->index;
if (*list != bfd_abs_section_ptr)
{
#define PREV_SEC(sec) (htab->stub_group[(sec)->id].link_sec)
which is what we want. */
PREV_SEC (isec) = *list;
*list = isec;
}
}
}
sections may result in fewer stubs. More importantly, we need to
put all .init* and .fini* stubs at the beginning of the .init or
.fini output sections respectively, because glibc splits the
_init and _fini functions into multiple parts. Putting a stub in
the middle of a function is not a good idea. */
static void
group_sections (struct elf32_arm_link_hash_table *htab,
bfd_size_type stub_group_size,
bfd_boolean stubs_always_before_branch)
{
asection **list = htab->input_list + htab->top_index;
do
{
asection *tail = *list;
if (tail == bfd_abs_section_ptr)
continue;
while (tail != NULL)
{
asection *curr;
asection *prev;
bfd_size_type total;
curr = tail;
total = tail->size;
while ((prev = PREV_SEC (curr)) != NULL
&& ((total += curr->output_offset - prev->output_offset)
< stub_group_size))
curr = prev;
than stub_group_size and thus can be handled by one stub
section. (Or the tail section is itself larger than
stub_group_size, in which case we may be toast.)
We should really be keeping track of the total size of
stubs added here, as stubs contribute to the final output
section size. */
do
{
prev = PREV_SEC (tail);
htab->stub_group[tail->id].link_sec = curr;
}
while (tail != curr && (tail = prev) != NULL);
bytes before the stub section can be handled by it too. */
if (!stubs_always_before_branch)
{
total = 0;
while (prev != NULL
&& ((total += tail->output_offset - prev->output_offset)
< stub_group_size))
{
tail = prev;
prev = PREV_SEC (tail);
htab->stub_group[tail->id].link_sec = curr;
}
}
tail = prev;
}
}
while (list-- != htab->input_list);
free (htab->input_list);
#undef PREV_SEC
}
The basic idea here is to examine all the relocations looking for
PC-relative calls to a target that is unreachable with a "bl"
instruction. */
bfd_boolean
elf32_arm_size_stubs (bfd *output_bfd,
bfd *stub_bfd,
struct bfd_link_info *info,
bfd_signed_vma group_size,
asection * (*add_stub_section) (const char *, asection *),
void (*layout_sections_again) (void))
{
bfd_size_type stub_group_size;
bfd_boolean stubs_always_before_branch;
bfd_boolean stub_changed = 0;
struct elf32_arm_link_hash_table *htab = elf32_arm_hash_table (info);
finalized when we created stub_bfd. */
bfd_set_arch_mach (stub_bfd, bfd_get_arch (output_bfd),
bfd_get_mach (output_bfd));
htab->stub_bfd = stub_bfd;
htab->add_stub_section = add_stub_section;
htab->layout_sections_again = layout_sections_again;
stubs_always_before_branch = group_size < 0;
if (group_size < 0)
stub_group_size = -group_size;
else
stub_group_size = group_size;
if (stub_group_size == 1)
{
maximum size (a given section can contain both ARM and Thumb
code, so the worst case has to be taken into account).
This value is 24K less than that, which allows for 2025
12-byte stubs. If we exceed that, then we will fail to link.
The user will have to relink with an explicit group size
option. */
stub_group_size = 4170000;
}
group_sections (htab, stub_group_size, stubs_always_before_branch);
while (1)
{
bfd *input_bfd;
unsigned int bfd_indx;
asection *stub_sec;
for (input_bfd = info->input_bfds, bfd_indx = 0;
input_bfd != NULL;
input_bfd = input_bfd->link_next, bfd_indx++)
{
Elf_Internal_Shdr *symtab_hdr;
asection *section;
Elf_Internal_Sym *local_syms = NULL;
symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
if (symtab_hdr->sh_info == 0)
continue;
for (section = input_bfd->sections;
section != NULL;
section = section->next)
{
Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
to do. */
if ((section->flags & SEC_RELOC) == 0
|| section->reloc_count == 0
|| (section->flags & SEC_CODE) == 0)
continue;
discarded, then don't create any stubs. */
if (section->output_section == NULL
|| section->output_section->owner != output_bfd)
continue;
internal_relocs
= _bfd_elf_link_read_relocs (input_bfd, section, NULL,
NULL, info->keep_memory);
if (internal_relocs == NULL)
goto error_ret_free_local;
irela = internal_relocs;
irelaend = irela + section->reloc_count;
for (; irela < irelaend; irela++)
{
unsigned int r_type, r_indx;
enum elf32_arm_stub_type stub_type;
struct elf32_arm_stub_hash_entry *stub_entry;
asection *sym_sec;
bfd_vma sym_value;
bfd_vma destination;
struct elf32_arm_link_hash_entry *hash;
const char *sym_name;
char *stub_name;
const asection *id_sec;
unsigned char st_type;
r_type = ELF32_R_TYPE (irela->r_info);
r_indx = ELF32_R_SYM (irela->r_info);
if (r_type >= (unsigned int) R_ARM_max)
{
bfd_set_error (bfd_error_bad_value);
error_ret_free_internal:
if (elf_section_data (section)->relocs == NULL)
free (internal_relocs);
goto error_ret_free_local;
}
if ((r_type != (unsigned int) R_ARM_CALL)
&& (r_type != (unsigned int) R_ARM_THM_CALL))
continue;
section. */
sym_sec = NULL;
sym_value = 0;
destination = 0;
hash = NULL;
sym_name = NULL;
if (r_indx < symtab_hdr->sh_info)
{
Elf_Internal_Sym *sym;
Elf_Internal_Shdr *hdr;
if (local_syms == NULL)
{
local_syms
= (Elf_Internal_Sym *) symtab_hdr->contents;
if (local_syms == NULL)
local_syms
= bfd_elf_get_elf_syms (input_bfd, symtab_hdr,
symtab_hdr->sh_info, 0,
NULL, NULL, NULL);
if (local_syms == NULL)
goto error_ret_free_internal;
}
sym = local_syms + r_indx;
hdr = elf_elfsections (input_bfd)[sym->st_shndx];
sym_sec = hdr->bfd_section;
if (ELF_ST_TYPE (sym->st_info) != STT_SECTION)
sym_value = sym->st_value;
destination = (sym_value + irela->r_addend
+ sym_sec->output_offset
+ sym_sec->output_section->vma);
st_type = ELF_ST_TYPE (sym->st_info);
sym_name
= bfd_elf_string_from_elf_section (input_bfd,
symtab_hdr->sh_link,
sym->st_name);
}
else
{
int e_indx;
e_indx = r_indx - symtab_hdr->sh_info;
hash = ((struct elf32_arm_link_hash_entry *)
elf_sym_hashes (input_bfd)[e_indx]);
while (hash->root.root.type == bfd_link_hash_indirect
|| hash->root.root.type == bfd_link_hash_warning)
hash = ((struct elf32_arm_link_hash_entry *)
hash->root.root.u.i.link);
if (hash->root.root.type == bfd_link_hash_defined
|| hash->root.root.type == bfd_link_hash_defweak)
{
sym_sec = hash->root.root.u.def.section;
sym_value = hash->root.root.u.def.value;
if (sym_sec->output_section != NULL)
destination = (sym_value + irela->r_addend
+ sym_sec->output_offset
+ sym_sec->output_section->vma);
}
else if (hash->root.root.type == bfd_link_hash_undefweak
|| hash->root.root.type == bfd_link_hash_undefined)
which is treated separately.
For absolute code, they cannot be handled. */
continue;
else
{
bfd_set_error (bfd_error_bad_value);
goto error_ret_free_internal;
}
st_type = ELF_ST_TYPE (hash->root.type);
sym_name = hash->root.root.root.string;
}
stub_type = arm_type_of_stub (info, section, irela, st_type,
hash, destination, sym_sec,
input_bfd, sym_name);
if (stub_type == arm_stub_none)
continue;
id_sec = htab->stub_group[section->id].link_sec;
stub_name = elf32_arm_stub_name (id_sec, sym_sec, hash, irela);
if (!stub_name)
goto error_ret_free_internal;
stub_entry = arm_stub_hash_lookup (&htab->stub_hash_table,
stub_name,
FALSE, FALSE);
if (stub_entry != NULL)
{
free (stub_name);
continue;
}
stub_entry = elf32_arm_add_stub (stub_name, section, htab);
if (stub_entry == NULL)
{
free (stub_name);
goto error_ret_free_internal;
}
stub_entry->target_value = sym_value;
stub_entry->target_section = sym_sec;
stub_entry->stub_type = stub_type;
stub_entry->h = hash;
stub_entry->st_type = st_type;
if (sym_name == NULL)
sym_name = "unnamed";
stub_entry->output_name
= bfd_alloc (htab->stub_bfd,
sizeof (THUMB2ARM_GLUE_ENTRY_NAME)
+ strlen (sym_name));
if (stub_entry->output_name == NULL)
{
free (stub_name);
goto error_ret_free_internal;
}
ARM-to-Thumb and Thumb-to-ARM stubs. */
if (r_type == (unsigned int) R_ARM_THM_CALL
&& st_type != STT_ARM_TFUNC)
sprintf (stub_entry->output_name, THUMB2ARM_GLUE_ENTRY_NAME,
sym_name);
else if (r_type == (unsigned int) R_ARM_CALL
&& st_type == STT_ARM_TFUNC)
sprintf (stub_entry->output_name, ARM2THUMB_GLUE_ENTRY_NAME,
sym_name);
else
sprintf (stub_entry->output_name, STUB_ENTRY_NAME,
sym_name);
stub_changed = TRUE;
}
if (elf_section_data (section)->relocs == NULL)
free (internal_relocs);
}
}
if (!stub_changed)
break;
stub sections. */
for (stub_sec = htab->stub_bfd->sections;
stub_sec != NULL;
stub_sec = stub_sec->next)
stub_sec->size = 0;
bfd_hash_traverse (&htab->stub_hash_table, arm_size_one_stub, htab);
(*htab->layout_sections_again) ();
stub_changed = FALSE;
}
return TRUE;
error_ret_free_local:
return FALSE;
}
stubs are kept in a hash table attached to the main linker hash
table. We also set up the .plt entries for statically linked PIC
functions here. This function is called via arm_elf_finish in the
linker. */
bfd_boolean
elf32_arm_build_stubs (struct bfd_link_info *info)
{
asection *stub_sec;
struct bfd_hash_table *table;
struct elf32_arm_link_hash_table *htab;
htab = elf32_arm_hash_table (info);
for (stub_sec = htab->stub_bfd->sections;
stub_sec != NULL;
stub_sec = stub_sec->next)
{
bfd_size_type size;
if (!strstr (stub_sec->name, STUB_SUFFIX))
continue;
size = stub_sec->size;
stub_sec->contents = bfd_zalloc (htab->stub_bfd, size);
if (stub_sec->contents == NULL && size != 0)
return FALSE;
stub_sec->size = 0;
}
table = &htab->stub_hash_table;
bfd_hash_traverse (table, arm_build_one_stub, info);
return TRUE;
}
static struct elf_link_hash_entry *
find_thumb_glue (struct bfd_link_info *link_info,
const char *name,
char **error_message)
{
char *tmp_name;
struct elf_link_hash_entry *hash;
struct elf32_arm_link_hash_table *hash_table;
hash_table = elf32_arm_hash_table (link_info);
tmp_name = bfd_malloc ((bfd_size_type) strlen (name)
+ strlen (THUMB2ARM_GLUE_ENTRY_NAME) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, THUMB2ARM_GLUE_ENTRY_NAME, name);
hash = elf_link_hash_lookup
(&(hash_table)->root, tmp_name, FALSE, FALSE, TRUE);
if (hash == NULL
&& asprintf (error_message, _("unable to find THUMB glue '%s' for '%s'"),
tmp_name, name) == -1)
*error_message = (char *) bfd_errmsg (bfd_error_system_call);
free (tmp_name);
return hash;
}
static struct elf_link_hash_entry *
find_arm_glue (struct bfd_link_info *link_info,
const char *name,
char **error_message)
{
char *tmp_name;
struct elf_link_hash_entry *myh;
struct elf32_arm_link_hash_table *hash_table;
hash_table = elf32_arm_hash_table (link_info);
tmp_name = bfd_malloc ((bfd_size_type) strlen (name)
+ strlen (ARM2THUMB_GLUE_ENTRY_NAME) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, ARM2THUMB_GLUE_ENTRY_NAME, name);
myh = elf_link_hash_lookup
(&(hash_table)->root, tmp_name, FALSE, FALSE, TRUE);
if (myh == NULL
&& asprintf (error_message, _("unable to find ARM glue '%s' for '%s'"),
tmp_name, name) == -1)
*error_message = (char *) bfd_errmsg (bfd_error_system_call);
free (tmp_name);
return myh;
}
.arm
__func_from_arm:
ldr r12, __func_addr
bx r12
__func_addr:
.word func @ behave as if you saw a ARM_32 reloc.
(v5t static images)
.arm
__func_from_arm:
ldr pc, __func_addr
__func_addr:
.word func @ behave as if you saw a ARM_32 reloc.
(relocatable images)
.arm
__func_from_arm:
ldr r12, __func_offset
add r12, r12, pc
bx r12
__func_offset:
.word func - . */
#define ARM2THUMB_STATIC_GLUE_SIZE 12
static const insn32 a2t1_ldr_insn = 0xe59fc000;
static const insn32 a2t2_bx_r12_insn = 0xe12fff1c;
static const insn32 a2t3_func_addr_insn = 0x00000001;
#define ARM2THUMB_V5_STATIC_GLUE_SIZE 8
static const insn32 a2t1v5_ldr_insn = 0xe51ff004;
static const insn32 a2t2v5_func_addr_insn = 0x00000001;
#define ARM2THUMB_PIC_GLUE_SIZE 16
static const insn32 a2t1p_ldr_insn = 0xe59fc004;
static const insn32 a2t2p_add_pc_insn = 0xe08cc00f;
static const insn32 a2t3p_bx_r12_insn = 0xe12fff1c;
.thumb .thumb
.align 2 .align 2
__func_from_thumb: __func_from_thumb:
bx pc push {r6, lr}
nop ldr r6, __func_addr
.arm mov lr, pc
b func bx r6
.arm
;; back_to_thumb
ldmia r13! {r6, lr}
bx lr
__func_addr:
.word func */
#define THUMB2ARM_GLUE_SIZE 8
static const insn16 t2a1_bx_pc_insn = 0x4778;
static const insn16 t2a2_noop_insn = 0x46c0;
static const insn32 t2a3_b_insn = 0xea000000;
#define VFP11_ERRATUM_VENEER_SIZE 8
#define ARM_BX_VENEER_SIZE 12
static const insn32 armbx1_tst_insn = 0xe3100001;
static const insn32 armbx2_moveq_insn = 0x01a0f000;
static const insn32 armbx3_bx_insn = 0xe12fff10;
#ifndef ELFARM_NABI_C_INCLUDED
static void
arm_allocate_glue_section_space (bfd * abfd, bfd_size_type size, const char * name)
{
asection * s;
bfd_byte * contents;
if (size == 0)
return;
BFD_ASSERT (abfd != NULL);
s = bfd_get_section_by_name (abfd, name);
BFD_ASSERT (s != NULL);
contents = bfd_alloc (abfd, size);
BFD_ASSERT (s->size == size);
s->contents = contents;
}
bfd_boolean
bfd_elf32_arm_allocate_interworking_sections (struct bfd_link_info * info)
{
struct elf32_arm_link_hash_table * globals;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
arm_allocate_glue_section_space (globals->bfd_of_glue_owner,
globals->arm_glue_size,
ARM2THUMB_GLUE_SECTION_NAME);
arm_allocate_glue_section_space (globals->bfd_of_glue_owner,
globals->thumb_glue_size,
THUMB2ARM_GLUE_SECTION_NAME);
arm_allocate_glue_section_space (globals->bfd_of_glue_owner,
globals->vfp11_erratum_glue_size,
VFP11_ERRATUM_VENEER_SECTION_NAME);
arm_allocate_glue_section_space (globals->bfd_of_glue_owner,
globals->bx_glue_size,
ARM_BX_GLUE_SECTION_NAME);
return TRUE;
}
returns the symbol identifying the stub. */
static struct elf_link_hash_entry *
record_arm_to_thumb_glue (struct bfd_link_info * link_info,
struct elf_link_hash_entry * h)
{
const char * name = h->root.root.string;
asection * s;
char * tmp_name;
struct elf_link_hash_entry * myh;
struct bfd_link_hash_entry * bh;
struct elf32_arm_link_hash_table * globals;
bfd_vma val;
bfd_size_type size;
globals = elf32_arm_hash_table (link_info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name
(globals->bfd_of_glue_owner, ARM2THUMB_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
tmp_name = bfd_malloc ((bfd_size_type) strlen (name) + strlen (ARM2THUMB_GLUE_ENTRY_NAME) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, ARM2THUMB_GLUE_ENTRY_NAME, name);
myh = elf_link_hash_lookup
(&(globals)->root, tmp_name, FALSE, FALSE, TRUE);
if (myh != NULL)
{
free (tmp_name);
return myh;
}
Even though the section isn't allocated yet, this is where we will be
putting it. The +1 on the value marks that the stub has not been
output yet - not that it is a Thumb function. */
bh = NULL;
val = globals->arm_glue_size + 1;
_bfd_generic_link_add_one_symbol (link_info, globals->bfd_of_glue_owner,
tmp_name, BSF_GLOBAL, s, val,
NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
myh->forced_local = 1;
free (tmp_name);
if (link_info->shared || globals->root.is_relocatable_executable
|| globals->pic_veneer)
size = ARM2THUMB_PIC_GLUE_SIZE;
else if (globals->use_blx)
size = ARM2THUMB_V5_STATIC_GLUE_SIZE;
else
size = ARM2THUMB_STATIC_GLUE_SIZE;
s->size += size;
globals->arm_glue_size += size;
return myh;
}
static void
record_thumb_to_arm_glue (struct bfd_link_info *link_info,
struct elf_link_hash_entry *h)
{
const char *name = h->root.root.string;
asection *s;
char *tmp_name;
struct elf_link_hash_entry *myh;
struct bfd_link_hash_entry *bh;
struct elf32_arm_link_hash_table *hash_table;
bfd_vma val;
hash_table = elf32_arm_hash_table (link_info);
BFD_ASSERT (hash_table != NULL);
BFD_ASSERT (hash_table->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name
(hash_table->bfd_of_glue_owner, THUMB2ARM_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
tmp_name = bfd_malloc ((bfd_size_type) strlen (name)
+ strlen (THUMB2ARM_GLUE_ENTRY_NAME) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, THUMB2ARM_GLUE_ENTRY_NAME, name);
myh = elf_link_hash_lookup
(&(hash_table)->root, tmp_name, FALSE, FALSE, TRUE);
if (myh != NULL)
{
free (tmp_name);
return;
}
Even though the section isn't allocated yet, this is where we will be
putting it. The +1 on the value marks that the stub has not been
output yet - not that it is a Thumb function. */
bh = NULL;
val = hash_table->thumb_glue_size + 1;
_bfd_generic_link_add_one_symbol (link_info, hash_table->bfd_of_glue_owner,
tmp_name, BSF_GLOBAL, s, val,
NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_ARM_TFUNC);
myh->forced_local = 1;
free (tmp_name);
#define CHANGE_TO_ARM "__%s_change_to_arm"
#define BACK_FROM_ARM "__%s_back_from_arm"
tmp_name = bfd_malloc ((bfd_size_type) strlen (name)
+ strlen (CHANGE_TO_ARM) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, CHANGE_TO_ARM, name);
bh = NULL;
val = hash_table->thumb_glue_size + 4,
_bfd_generic_link_add_one_symbol (link_info, hash_table->bfd_of_glue_owner,
tmp_name, BSF_LOCAL, s, val,
NULL, TRUE, FALSE, &bh);
free (tmp_name);
s->size += THUMB2ARM_GLUE_SIZE;
hash_table->thumb_glue_size += THUMB2ARM_GLUE_SIZE;
}
static void
record_arm_bx_glue (struct bfd_link_info * link_info, int reg)
{
asection * s;
struct elf32_arm_link_hash_table *globals;
char *tmp_name;
struct elf_link_hash_entry *myh;
struct bfd_link_hash_entry *bh;
bfd_vma val;
if (reg == 15)
return;
globals = elf32_arm_hash_table (link_info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
if (globals->bx_glue_offset[reg])
return;
s = bfd_get_section_by_name
(globals->bfd_of_glue_owner, ARM_BX_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
tmp_name = bfd_malloc ((bfd_size_type) strlen (ARM_BX_GLUE_ENTRY_NAME) + 1);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, ARM_BX_GLUE_ENTRY_NAME, reg);
myh = elf_link_hash_lookup
(&(globals)->root, tmp_name, FALSE, FALSE, FALSE);
BFD_ASSERT (myh == NULL);
bh = NULL;
val = globals->bx_glue_size;
_bfd_generic_link_add_one_symbol (link_info, globals->bfd_of_glue_owner,
tmp_name, BSF_FUNCTION | BSF_LOCAL, s, val,
NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
myh->forced_local = 1;
s->size += ARM_BX_VENEER_SIZE;
globals->bx_glue_offset[reg] = globals->bx_glue_size | 2;
globals->bx_glue_size += ARM_BX_VENEER_SIZE;
}
static void
elf32_arm_section_map_add (asection *sec, char type, bfd_vma vma)
{
struct _arm_elf_section_data *sec_data = elf32_arm_section_data (sec);
unsigned int newidx;
if (sec_data->map == NULL)
{
sec_data->map = bfd_malloc (sizeof (elf32_arm_section_map));
sec_data->mapcount = 0;
sec_data->mapsize = 1;
}
newidx = sec_data->mapcount++;
if (sec_data->mapcount > sec_data->mapsize)
{
sec_data->mapsize *= 2;
sec_data->map = bfd_realloc_or_free (sec_data->map, sec_data->mapsize
* sizeof (elf32_arm_section_map));
}
if (sec_data->map)
{
sec_data->map[newidx].vma = vma;
sec_data->map[newidx].type = type;
}
}
veneers are handled for now. */
static bfd_vma
record_vfp11_erratum_veneer (struct bfd_link_info *link_info,
elf32_vfp11_erratum_list *branch,
bfd *branch_bfd,
asection *branch_sec,
unsigned int offset)
{
asection *s;
struct elf32_arm_link_hash_table *hash_table;
char *tmp_name;
struct elf_link_hash_entry *myh;
struct bfd_link_hash_entry *bh;
bfd_vma val;
struct _arm_elf_section_data *sec_data;
int errcount;
elf32_vfp11_erratum_list *newerr;
hash_table = elf32_arm_hash_table (link_info);
BFD_ASSERT (hash_table != NULL);
BFD_ASSERT (hash_table->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name
(hash_table->bfd_of_glue_owner, VFP11_ERRATUM_VENEER_SECTION_NAME);
sec_data = elf32_arm_section_data (s);
BFD_ASSERT (s != NULL);
tmp_name = bfd_malloc ((bfd_size_type) strlen
(VFP11_ERRATUM_VENEER_ENTRY_NAME) + 10);
BFD_ASSERT (tmp_name);
sprintf (tmp_name, VFP11_ERRATUM_VENEER_ENTRY_NAME,
hash_table->num_vfp11_fixes);
myh = elf_link_hash_lookup
(&(hash_table)->root, tmp_name, FALSE, FALSE, FALSE);
BFD_ASSERT (myh == NULL);
bh = NULL;
val = hash_table->vfp11_erratum_glue_size;
_bfd_generic_link_add_one_symbol (link_info, hash_table->bfd_of_glue_owner,
tmp_name, BSF_FUNCTION | BSF_LOCAL, s, val,
NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
myh->forced_local = 1;
errcount = ++(sec_data->erratumcount);
newerr = bfd_zmalloc (sizeof (elf32_vfp11_erratum_list));
newerr->type = VFP11_ERRATUM_ARM_VENEER;
newerr->vma = -1;
newerr->u.v.branch = branch;
newerr->u.v.id = hash_table->num_vfp11_fixes;
branch->u.b.veneer = newerr;
newerr->next = sec_data->erratumlist;
sec_data->erratumlist = newerr;
sprintf (tmp_name, VFP11_ERRATUM_VENEER_ENTRY_NAME "_r",
hash_table->num_vfp11_fixes);
myh = elf_link_hash_lookup
(&(hash_table)->root, tmp_name, FALSE, FALSE, FALSE);
if (myh != NULL)
abort ();
bh = NULL;
val = offset + 4;
_bfd_generic_link_add_one_symbol (link_info, branch_bfd, tmp_name, BSF_LOCAL,
branch_sec, val, NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
myh->forced_local = 1;
free (tmp_name);
entry for that symbol to the code/data map for the section. */
if (hash_table->vfp11_erratum_glue_size == 0)
{
bh = NULL;
ever requires this erratum fix. */
_bfd_generic_link_add_one_symbol (link_info,
hash_table->bfd_of_glue_owner, "$a",
BSF_LOCAL, s, 0, NULL,
TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_NOTYPE);
myh->forced_local = 1;
BFDs. We must make a note of this generated mapping symbol
ourselves so that code byteswapping works properly in
elf32_arm_write_section. */
elf32_arm_section_map_add (s, 'a', 0);
}
s->size += VFP11_ERRATUM_VENEER_SIZE;
hash_table->vfp11_erratum_glue_size += VFP11_ERRATUM_VENEER_SIZE;
hash_table->num_vfp11_fixes++;
return val;
}
would prevent elf_link_input_bfd() from processing the contents
of the section. */
#define ARM_GLUE_SECTION_FLAGS \
(SEC_ALLOC | SEC_LOAD | SEC_HAS_CONTENTS | SEC_IN_MEMORY | SEC_CODE | SEC_READONLY)
static bfd_boolean
arm_make_glue_section (bfd * abfd, const char * name)
{
asection * sec;
sec = bfd_get_section_by_name (abfd, name);
if (sec != NULL)
return TRUE;
sec = bfd_make_section_with_flags (abfd, name, ARM_GLUE_SECTION_FLAGS);
if (sec == NULL
|| !bfd_set_section_alignment (abfd, sec, 2))
return FALSE;
collection, despite the fact that no relocs refer to this section. */
sec->gc_mark = 1;
return TRUE;
}
linker scripts in ld/emultempl/{armelf}.em. */
bfd_boolean
bfd_elf32_arm_add_glue_sections_to_bfd (bfd *abfd,
struct bfd_link_info *info)
{
link do not bother adding the glue. */
if (info->relocatable)
return TRUE;
if (!strcmp (abfd->filename, "linker stubs"))
return TRUE;
return arm_make_glue_section (abfd, ARM2THUMB_GLUE_SECTION_NAME)
&& arm_make_glue_section (abfd, THUMB2ARM_GLUE_SECTION_NAME)
&& arm_make_glue_section (abfd, VFP11_ERRATUM_VENEER_SECTION_NAME)
&& arm_make_glue_section (abfd, ARM_BX_GLUE_SECTION_NAME);
}
This function is called from the linker scripts in ld/emultempl/
{armelf/pe}.em. */
bfd_boolean
bfd_elf32_arm_get_bfd_for_interworking (bfd *abfd, struct bfd_link_info *info)
{
struct elf32_arm_link_hash_table *globals;
do not bother getting a bfd to hold the glue. */
if (info->relocatable)
return TRUE;
BFD_ASSERT (!(abfd->flags & DYNAMIC));
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
if (globals->bfd_of_glue_owner != NULL)
return TRUE;
globals->bfd_of_glue_owner = abfd;
return TRUE;
}
static void
check_use_blx (struct elf32_arm_link_hash_table *globals)
{
if (bfd_elf_get_obj_attr_int (globals->obfd, OBJ_ATTR_PROC,
Tag_CPU_arch) > 2)
globals->use_blx = 1;
}
bfd_boolean
bfd_elf32_arm_process_before_allocation (bfd *abfd,
struct bfd_link_info *link_info)
{
Elf_Internal_Shdr *symtab_hdr;
Elf_Internal_Rela *internal_relocs = NULL;
Elf_Internal_Rela *irel, *irelend;
bfd_byte *contents = NULL;
asection *sec;
struct elf32_arm_link_hash_table *globals;
to construct any glue. */
if (link_info->relocatable)
return TRUE;
hook to do reloc rummaging, before section sizes are nailed down. */
globals = elf32_arm_hash_table (link_info);
BFD_ASSERT (globals != NULL);
check_use_blx (globals);
if (globals->byteswap_code && !bfd_big_endian (abfd))
{
_bfd_error_handler (_("%B: BE8 images only valid in big-endian mode."),
abfd);
return FALSE;
}
the output then we will not have a glue owner bfd. This is OK, it
just means that there is nothing else for us to do here. */
if (globals->bfd_of_glue_owner == NULL)
return TRUE;
sec = abfd->sections;
if (sec == NULL)
return TRUE;
for (; sec != NULL; sec = sec->next)
{
if (sec->reloc_count == 0)
continue;
if ((sec->flags & SEC_EXCLUDE) != 0)
continue;
symtab_hdr = & elf_symtab_hdr (abfd);
internal_relocs
= _bfd_elf_link_read_relocs (abfd, sec, NULL, NULL, FALSE);
if (internal_relocs == NULL)
goto error_return;
irelend = internal_relocs + sec->reloc_count;
for (irel = internal_relocs; irel < irelend; irel++)
{
long r_type;
unsigned long r_index;
struct elf_link_hash_entry *h;
r_type = ELF32_R_TYPE (irel->r_info);
r_index = ELF32_R_SYM (irel->r_info);
if ( r_type != R_ARM_PC24
&& r_type != R_ARM_PLT32
&& r_type != R_ARM_JUMP24
&& r_type != R_ARM_THM_JUMP24
&& (r_type != R_ARM_V4BX || globals->fix_v4bx < 2))
continue;
if (contents == NULL)
{
if (elf_section_data (sec)->this_hdr.contents != NULL)
contents = elf_section_data (sec)->this_hdr.contents;
else
{
if (! bfd_malloc_and_get_section (abfd, sec, &contents))
goto error_return;
}
}
if (r_type == R_ARM_V4BX)
{
int reg;
reg = bfd_get_32 (abfd, contents + irel->r_offset) & 0xf;
record_arm_bx_glue (link_info, reg);
continue;
}
h = NULL;
if (r_index < symtab_hdr->sh_info)
continue;
r_index -= symtab_hdr->sh_info;
h = (struct elf_link_hash_entry *)
elf_sym_hashes (abfd)[r_index];
the current section and so cannot be a cross ARM/Thumb relocation. */
if (h == NULL)
continue;
glue. */
if (globals->splt != NULL && h->plt.offset != (bfd_vma) -1)
continue;
switch (r_type)
{
case R_ARM_PC24:
case R_ARM_PLT32:
case R_ARM_JUMP24:
the target of the call. If it is a thumb target, we
insert glue. */
if (ELF_ST_TYPE (h->type) == STT_ARM_TFUNC
&& !(r_type == R_ARM_CALL && globals->use_blx))
record_arm_to_thumb_glue (link_info, h);
break;
case R_ARM_THM_JUMP24:
up the target of the call. If it is not a thumb
target, we insert glue. */
if (ELF_ST_TYPE (h->type) != STT_ARM_TFUNC
&& !(globals->use_blx && r_type == R_ARM_THM_CALL)
&& h->root.type != bfd_link_hash_undefweak)
record_thumb_to_arm_glue (link_info, h);
break;
default:
abort ();
}
}
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
free (contents);
contents = NULL;
if (internal_relocs != NULL
&& elf_section_data (sec)->relocs != internal_relocs)
free (internal_relocs);
internal_relocs = NULL;
}
return TRUE;
error_return:
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
free (contents);
if (internal_relocs != NULL
&& elf_section_data (sec)->relocs != internal_relocs)
free (internal_relocs);
return FALSE;
}
#endif
void
bfd_elf32_arm_init_maps (bfd *abfd)
{
Elf_Internal_Sym *isymbuf;
Elf_Internal_Shdr *hdr;
unsigned int i, localsyms;
if ((abfd->flags & DYNAMIC) != 0)
return;
hdr = & elf_symtab_hdr (abfd);
localsyms = hdr->sh_info;
should contain the number of local symbols, which should come before any
global symbols. Mapping symbols are always local. */
isymbuf = bfd_elf_get_elf_syms (abfd, hdr, localsyms, 0, NULL, NULL,
NULL);
if (isymbuf == NULL)
return;
for (i = 0; i < localsyms; i++)
{
Elf_Internal_Sym *isym = &isymbuf[i];
asection *sec = bfd_section_from_elf_index (abfd, isym->st_shndx);
const char *name;
if (sec != NULL
&& ELF_ST_BIND (isym->st_info) == STB_LOCAL)
{
name = bfd_elf_string_from_elf_section (abfd,
hdr->sh_link, isym->st_name);
if (bfd_is_arm_special_symbol_name (name,
BFD_ARM_SPECIAL_SYM_TYPE_MAP))
elf32_arm_section_map_add (sec, name[1], isym->st_value);
}
}
}
void
bfd_elf32_arm_set_vfp11_fix (bfd *obfd, struct bfd_link_info *link_info)
{
struct elf32_arm_link_hash_table *globals = elf32_arm_hash_table (link_info);
obj_attribute *out_attr = elf_known_obj_attributes_proc (obfd);
if (out_attr[Tag_CPU_arch].i >= TAG_CPU_ARCH_V7)
{
switch (globals->vfp11_fix)
{
case BFD_ARM_VFP11_FIX_DEFAULT:
case BFD_ARM_VFP11_FIX_NONE:
globals->vfp11_fix = BFD_ARM_VFP11_FIX_NONE;
break;
default:
(*_bfd_error_handler) (_("%B: warning: selected VFP11 erratum "
"workaround is not necessary for target architecture"), obfd);
}
}
else if (globals->vfp11_fix == BFD_ARM_VFP11_FIX_DEFAULT)
enable it by default. If users is running with broken hardware, they
must enable the erratum fix explicitly. */
globals->vfp11_fix = BFD_ARM_VFP11_FIX_NONE;
}
enum bfd_arm_vfp11_pipe
{
VFP11_FMAC,
VFP11_LS,
VFP11_DS,
VFP11_BAD
};
registers, or X:RX for double-precision registers, where RX is the group of
four bits in the instruction encoding and X is the single extension bit.
RX and X fields are specified using their lowest (starting) bit. The return
value is:
0...31: single-precision registers s0...s31
32...63: double-precision registers d0...d31.
Although X should be zero for VFP11 (encoding d0...d15 only), we might
encounter VFP3 instructions, so we allow the full range for DP registers. */
static unsigned int
bfd_arm_vfp11_regno (unsigned int insn, bfd_boolean is_double, unsigned int rx,
unsigned int x)
{
if (is_double)
return (((insn >> rx) & 0xf) | (((insn >> x) & 1) << 4)) + 32;
else
return (((insn >> rx) & 0xf) << 1) | ((insn >> x) & 1);
}
bfd_arm_vfp11_regno(). Ignore d16-d31. */
static void
bfd_arm_vfp11_write_mask (unsigned int *wmask, unsigned int reg)
{
if (reg < 32)
*wmask |= 1 << reg;
else if (reg < 48)
*wmask |= 3 << ((reg - 32) * 2);
}
static bfd_boolean
bfd_arm_vfp11_antidependency (unsigned int wmask, int *regs, int numregs)
{
int i;
for (i = 0; i < numregs; i++)
{
unsigned int reg = regs[i];
if (reg < 32 && (wmask & (1 << reg)) != 0)
return TRUE;
reg -= 32;
if (reg >= 16)
continue;
if ((wmask & (3 << (reg * 2))) != 0)
return TRUE;
}
return FALSE;
}
for VFP data-processing instructions, and finding the set of registers which
arbitrary VFP instructions may write to. We use a 32-bit unsigned int to
hold the written set, so FLDM etc. are easy to deal with (we're only
interested in 32 SP registers or 16 dp registers, due to the VFP version
implemented by the chip in question). DP registers are marked by setting
both SP registers in the write mask). */
static enum bfd_arm_vfp11_pipe
bfd_arm_vfp11_insn_decode (unsigned int insn, unsigned int *destmask, int *regs,
int *numregs)
{
enum bfd_arm_vfp11_pipe pipe = VFP11_BAD;
bfd_boolean is_double = ((insn & 0xf00) == 0xb00) ? 1 : 0;
if ((insn & 0x0f000e10) == 0x0e000a00)
{
unsigned int pqrs;
unsigned int fd = bfd_arm_vfp11_regno (insn, is_double, 12, 22);
unsigned int fm = bfd_arm_vfp11_regno (insn, is_double, 0, 5);
pqrs = ((insn & 0x00800000) >> 20)
| ((insn & 0x00300000) >> 19)
| ((insn & 0x00000040) >> 6);
switch (pqrs)
{
case 0:
case 1:
case 2:
case 3:
pipe = VFP11_FMAC;
bfd_arm_vfp11_write_mask (destmask, fd);
regs[0] = fd;
regs[1] = bfd_arm_vfp11_regno (insn, is_double, 16, 7);
regs[2] = fm;
*numregs = 3;
break;
case 4:
case 5:
case 6:
case 7:
pipe = VFP11_FMAC;
goto vfp_binop;
case 8:
pipe = VFP11_DS;
vfp_binop:
bfd_arm_vfp11_write_mask (destmask, fd);
regs[0] = bfd_arm_vfp11_regno (insn, is_double, 16, 7);
regs[1] = fm;
*numregs = 2;
break;
case 15:
{
unsigned int extn = ((insn >> 15) & 0x1e)
| ((insn >> 7) & 1);
switch (extn)
{
case 0:
case 1:
case 2:
case 8:
case 9:
case 10:
case 11:
case 16:
case 17:
case 24:
case 25:
case 26:
case 27:
*numregs = 0;
pipe = VFP11_FMAC;
break;
case 3:
registers to cause the erratum in previous instructions. */
bfd_arm_vfp11_write_mask (destmask, fd);
pipe = VFP11_DS;
break;
case 15:
{
int rnum = 0;
bfd_arm_vfp11_write_mask (destmask, fd);
if ((insn & 0x100) != 0)
regs[rnum++] = fm;
*numregs = rnum;
pipe = VFP11_FMAC;
}
break;
default:
return VFP11_BAD;
}
}
break;
default:
return VFP11_BAD;
}
}
else if ((insn & 0x0fe00ed0) == 0x0c400a10)
{
unsigned int fm = bfd_arm_vfp11_regno (insn, is_double, 0, 5);
if ((insn & 0x100000) == 0)
{
if (is_double)
bfd_arm_vfp11_write_mask (destmask, fm);
else
{
bfd_arm_vfp11_write_mask (destmask, fm);
bfd_arm_vfp11_write_mask (destmask, fm + 1);
}
}
pipe = VFP11_LS;
}
else if ((insn & 0x0e100e00) == 0x0c100a00)
{
int fd = bfd_arm_vfp11_regno (insn, is_double, 12, 22);
unsigned int puw = ((insn >> 21) & 0x1) | (((insn >> 23) & 3) << 1);
switch (puw)
{
case 0:
abort ();
case 2:
case 3:
case 5:
{
unsigned int i, offset = insn & 0xff;
if (is_double)
offset >>= 1;
for (i = fd; i < fd + offset; i++)
bfd_arm_vfp11_write_mask (destmask, i);
}
break;
case 4:
case 6:
bfd_arm_vfp11_write_mask (destmask, fd);
break;
default:
return VFP11_BAD;
}
pipe = VFP11_LS;
}
else if ((insn & 0x0f100e10) == 0x0e000a10)
{
unsigned int opcode = (insn >> 21) & 7;
unsigned int fn = bfd_arm_vfp11_regno (insn, is_double, 16, 7);
switch (opcode)
{
case 0:
case 1:
destination register. I don't know if this is exactly right,
but it is the conservative choice. */
bfd_arm_vfp11_write_mask (destmask, fn);
break;
case 7:
break;
}
pipe = VFP11_LS;
}
return pipe;
}
static int elf32_arm_compare_mapping (const void * a, const void * b);
VFP11 denormal/antidependency erratum. See, e.g., the ARM1136 errata sheet
(available from ARM) for details of the erratum. A short version is
described in ld.texinfo. */
bfd_boolean
bfd_elf32_arm_vfp11_erratum_scan (bfd *abfd, struct bfd_link_info *link_info)
{
asection *sec;
bfd_byte *contents = NULL;
int state = 0;
int regs[3], numregs = 0;
struct elf32_arm_link_hash_table *globals = elf32_arm_hash_table (link_info);
int use_vector = (globals->vfp11_fix == BFD_ARM_VFP11_FIX_VECTOR);
The states transition as follows:
0 -> 1 (vector) or 0 -> 2 (scalar)
A VFP FMAC-pipeline instruction has been seen. Fill
regs[0]..regs[numregs-1] with its input operands. Remember this
instruction in 'first_fmac'.
1 -> 2
Any instruction, except for a VFP instruction which overwrites
regs[*].
1 -> 3 [ -> 0 ] or
2 -> 3 [ -> 0 ]
A VFP instruction has been seen which overwrites any of regs[*].
We must make a veneer! Reset state to 0 before examining next
instruction.
2 -> 0
If we fail to match anything in state 2, reset to state 0 and reset
the instruction pointer to the instruction after 'first_fmac'.
If the VFP11 vector mode is in use, there must be at least two unrelated
instructions between anti-dependent VFP11 instructions to properly avoid
triggering the erratum, hence the use of the extra state 1. */
to construct any glue. */
if (link_info->relocatable)
return TRUE;
if (! is_arm_elf (abfd))
return TRUE;
BFD_ASSERT (globals->vfp11_fix != BFD_ARM_VFP11_FIX_DEFAULT);
if (globals->vfp11_fix == BFD_ARM_VFP11_FIX_NONE)
return TRUE;
for (sec = abfd->sections; sec != NULL; sec = sec->next)
{
unsigned int i, span, first_fmac = 0, veneer_of_insn = 0;
struct _arm_elf_section_data *sec_data;
section. Also skip if section is to be excluded. */
if (elf_section_type (sec) != SHT_PROGBITS
|| (elf_section_flags (sec) & SHF_EXECINSTR) == 0
|| (sec->flags & SEC_EXCLUDE) != 0
|| strcmp (sec->name, VFP11_ERRATUM_VENEER_SECTION_NAME) == 0)
continue;
sec_data = elf32_arm_section_data (sec);
if (sec_data->mapcount == 0)
continue;
if (elf_section_data (sec)->this_hdr.contents != NULL)
contents = elf_section_data (sec)->this_hdr.contents;
else if (! bfd_malloc_and_get_section (abfd, sec, &contents))
goto error_return;
qsort (sec_data->map, sec_data->mapcount, sizeof (elf32_arm_section_map),
elf32_arm_compare_mapping);
for (span = 0; span < sec_data->mapcount; span++)
{
unsigned int span_start = sec_data->map[span].vma;
unsigned int span_end = (span == sec_data->mapcount - 1)
? sec->size : sec_data->map[span + 1].vma;
char span_type = sec_data->map[span].type;
support Thumb-2 mode also at some point. */
if (span_type != 'a')
continue;
for (i = span_start; i < span_end;)
{
unsigned int next_i = i + 4;
unsigned int insn = bfd_big_endian (abfd)
? (contents[i] << 24)
| (contents[i + 1] << 16)
| (contents[i + 2] << 8)
| contents[i + 3]
: (contents[i + 3] << 24)
| (contents[i + 2] << 16)
| (contents[i + 1] << 8)
| contents[i];
unsigned int writemask = 0;
enum bfd_arm_vfp11_pipe pipe;
switch (state)
{
case 0:
pipe = bfd_arm_vfp11_insn_decode (insn, &writemask, regs,
&numregs);
operands on either the FMAC or the DS pipeline. This might
lead to slightly overenthusiastic veneer insertion. */
if (pipe == VFP11_FMAC || pipe == VFP11_DS)
{
state = use_vector ? 1 : 2;
first_fmac = i;
veneer_of_insn = insn;
}
break;
case 1:
{
int other_regs[3], other_numregs;
pipe = bfd_arm_vfp11_insn_decode (insn, &writemask,
other_regs,
&other_numregs);
if (pipe != VFP11_BAD
&& bfd_arm_vfp11_antidependency (writemask, regs,
numregs))
state = 3;
else
state = 2;
}
break;
case 2:
{
int other_regs[3], other_numregs;
pipe = bfd_arm_vfp11_insn_decode (insn, &writemask,
other_regs,
&other_numregs);
if (pipe != VFP11_BAD
&& bfd_arm_vfp11_antidependency (writemask, regs,
numregs))
state = 3;
else
{
state = 0;
next_i = first_fmac + 4;
}
}
break;
case 3:
abort ();
}
if (state == 3)
{
elf32_vfp11_erratum_list *newerr
= bfd_zmalloc (sizeof (elf32_vfp11_erratum_list));
int errcount;
errcount = ++(elf32_arm_section_data (sec)->erratumcount);
newerr->u.b.vfp_insn = veneer_of_insn;
switch (span_type)
{
case 'a':
newerr->type = VFP11_ERRATUM_BRANCH_TO_ARM_VENEER;
break;
default:
abort ();
}
record_vfp11_erratum_veneer (link_info, newerr, abfd, sec,
first_fmac);
newerr->vma = -1;
newerr->next = sec_data->erratumlist;
sec_data->erratumlist = newerr;
state = 0;
}
i = next_i;
}
}
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
free (contents);
contents = NULL;
}
return TRUE;
error_return:
if (contents != NULL
&& elf_section_data (sec)->this_hdr.contents != contents)
free (contents);
return FALSE;
}
after sections have been laid out, using specially-named symbols. */
void
bfd_elf32_arm_vfp11_fix_veneer_locations (bfd *abfd,
struct bfd_link_info *link_info)
{
asection *sec;
struct elf32_arm_link_hash_table *globals;
char *tmp_name;
if (link_info->relocatable)
return;
if (! is_arm_elf (abfd))
return;
globals = elf32_arm_hash_table (link_info);
tmp_name = bfd_malloc ((bfd_size_type) strlen
(VFP11_ERRATUM_VENEER_ENTRY_NAME) + 10);
for (sec = abfd->sections; sec != NULL; sec = sec->next)
{
struct _arm_elf_section_data *sec_data = elf32_arm_section_data (sec);
elf32_vfp11_erratum_list *errnode = sec_data->erratumlist;
for (; errnode != NULL; errnode = errnode->next)
{
struct elf_link_hash_entry *myh;
bfd_vma vma;
switch (errnode->type)
{
case VFP11_ERRATUM_BRANCH_TO_ARM_VENEER:
case VFP11_ERRATUM_BRANCH_TO_THUMB_VENEER:
sprintf (tmp_name, VFP11_ERRATUM_VENEER_ENTRY_NAME,
errnode->u.b.veneer->u.v.id);
myh = elf_link_hash_lookup
(&(globals)->root, tmp_name, FALSE, FALSE, TRUE);
if (myh == NULL)
(*_bfd_error_handler) (_("%B: unable to find VFP11 veneer "
"`%s'"), abfd, tmp_name);
vma = myh->root.u.def.section->output_section->vma
+ myh->root.u.def.section->output_offset
+ myh->root.u.def.value;
errnode->u.b.veneer->vma = vma;
break;
case VFP11_ERRATUM_ARM_VENEER:
case VFP11_ERRATUM_THUMB_VENEER:
sprintf (tmp_name, VFP11_ERRATUM_VENEER_ENTRY_NAME "_r",
errnode->u.v.id);
myh = elf_link_hash_lookup
(&(globals)->root, tmp_name, FALSE, FALSE, TRUE);
if (myh == NULL)
(*_bfd_error_handler) (_("%B: unable to find VFP11 veneer "
"`%s'"), abfd, tmp_name);
vma = myh->root.u.def.section->output_section->vma
+ myh->root.u.def.section->output_offset
+ myh->root.u.def.value;
errnode->u.v.branch->vma = vma;
break;
default:
abort ();
}
}
}
free (tmp_name);
}
void
bfd_elf32_arm_set_target_relocs (struct bfd *output_bfd,
struct bfd_link_info *link_info,
int target1_is_rel,
char * target2_type,
int fix_v4bx,
int use_blx,
bfd_arm_vfp11_fix vfp11_fix,
int no_enum_warn, int no_wchar_warn,
int pic_veneer)
{
struct elf32_arm_link_hash_table *globals;
globals = elf32_arm_hash_table (link_info);
globals->target1_is_rel = target1_is_rel;
if (strcmp (target2_type, "rel") == 0)
globals->target2_reloc = R_ARM_REL32;
else if (strcmp (target2_type, "abs") == 0)
globals->target2_reloc = R_ARM_ABS32;
else if (strcmp (target2_type, "got-rel") == 0)
globals->target2_reloc = R_ARM_GOT_PREL;
else
{
_bfd_error_handler (_("Invalid TARGET2 relocation type '%s'."),
target2_type);
}
globals->fix_v4bx = fix_v4bx;
globals->use_blx |= use_blx;
globals->vfp11_fix = vfp11_fix;
globals->pic_veneer = pic_veneer;
BFD_ASSERT (is_arm_elf (output_bfd));
elf_arm_tdata (output_bfd)->no_enum_size_warning = no_enum_warn;
elf_arm_tdata (output_bfd)->no_wchar_size_warning = no_wchar_warn;
}
static void
insert_thumb_branch (bfd *abfd, long int offset, bfd_byte *insn)
{
bfd_vma upper;
bfd_vma lower;
int reloc_sign;
BFD_ASSERT ((offset & 1) == 0);
upper = bfd_get_16 (abfd, insn);
lower = bfd_get_16 (abfd, insn + 2);
reloc_sign = (offset < 0) ? 1 : 0;
upper = (upper & ~(bfd_vma) 0x7ff)
| ((offset >> 12) & 0x3ff)
| (reloc_sign << 10);
lower = (lower & ~(bfd_vma) 0x2fff)
| (((!((offset >> 23) & 1)) ^ reloc_sign) << 13)
| (((!((offset >> 22) & 1)) ^ reloc_sign) << 11)
| ((offset >> 1) & 0x7ff);
bfd_put_16 (abfd, upper, insn);
bfd_put_16 (abfd, lower, insn + 2);
}
static int
elf32_thumb_to_arm_stub (struct bfd_link_info * info,
const char * name,
bfd * input_bfd,
bfd * output_bfd,
asection * input_section,
bfd_byte * hit_data,
asection * sym_sec,
bfd_vma offset,
bfd_signed_vma addend,
bfd_vma val,
char **error_message)
{
asection * s = 0;
bfd_vma my_offset;
long int ret_offset;
struct elf_link_hash_entry * myh;
struct elf32_arm_link_hash_table * globals;
myh = find_thumb_glue (info, name, error_message);
if (myh == NULL)
return FALSE;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
my_offset = myh->root.u.def.value;
s = bfd_get_section_by_name (globals->bfd_of_glue_owner,
THUMB2ARM_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
BFD_ASSERT (s->contents != NULL);
BFD_ASSERT (s->output_section != NULL);
if ((my_offset & 0x01) == 0x01)
{
if (sym_sec != NULL
&& sym_sec->owner != NULL
&& !INTERWORK_FLAG (sym_sec->owner))
{
(*_bfd_error_handler)
(_("%B(%s): warning: interworking not enabled.\n"
" first occurrence: %B: thumb call to arm"),
sym_sec->owner, input_bfd, name);
return FALSE;
}
--my_offset;
myh->root.u.def.value = my_offset;
put_thumb_insn (globals, output_bfd, (bfd_vma) t2a1_bx_pc_insn,
s->contents + my_offset);
put_thumb_insn (globals, output_bfd, (bfd_vma) t2a2_noop_insn,
s->contents + my_offset + 2);
ret_offset =
((bfd_signed_vma) val)
- ((bfd_signed_vma)
to the start of the stubs. */
(s->output_offset
+ my_offset
+ s->output_section->vma)
+ 4
+ 8);
put_arm_insn (globals, output_bfd,
(bfd_vma) t2a3_b_insn | ((ret_offset >> 2) & 0x00FFFFFF),
s->contents + my_offset + 4);
}
BFD_ASSERT (my_offset <= globals->thumb_glue_size);
ret_offset =
(s->output_section->vma + s->output_offset + my_offset)
- (input_section->output_section->vma + input_section->output_offset
+ offset)
- addend
- 8;
insert_thumb_branch (input_bfd, ret_offset, hit_data - input_section->vma);
return TRUE;
}
static struct elf_link_hash_entry *
elf32_arm_create_thumb_stub (struct bfd_link_info * info,
const char * name,
bfd * input_bfd,
bfd * output_bfd,
asection * sym_sec,
bfd_vma val,
asection * s,
char ** error_message)
{
bfd_vma my_offset;
long int ret_offset;
struct elf_link_hash_entry * myh;
struct elf32_arm_link_hash_table * globals;
myh = find_arm_glue (info, name, error_message);
if (myh == NULL)
return NULL;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
my_offset = myh->root.u.def.value;
if ((my_offset & 0x01) == 0x01)
{
if (sym_sec != NULL
&& sym_sec->owner != NULL
&& !INTERWORK_FLAG (sym_sec->owner))
{
(*_bfd_error_handler)
(_("%B(%s): warning: interworking not enabled.\n"
" first occurrence: %B: arm call to thumb"),
sym_sec->owner, input_bfd, name);
}
--my_offset;
myh->root.u.def.value = my_offset;
if (info->shared || globals->root.is_relocatable_executable
|| globals->pic_veneer)
{
so construct the address from a relative offset. */
constructing the address with adds. */
put_arm_insn (globals, output_bfd, (bfd_vma) a2t1p_ldr_insn,
s->contents + my_offset);
put_arm_insn (globals, output_bfd, (bfd_vma) a2t2p_add_pc_insn,
s->contents + my_offset + 4);
put_arm_insn (globals, output_bfd, (bfd_vma) a2t3p_bx_r12_insn,
s->contents + my_offset + 8);
and 8 for the pipeline offset. */
ret_offset = (val - (s->output_offset
+ s->output_section->vma
+ my_offset + 12))
| 1;
bfd_put_32 (output_bfd, ret_offset,
s->contents + my_offset + 12);
}
else if (globals->use_blx)
{
put_arm_insn (globals, output_bfd, (bfd_vma) a2t1v5_ldr_insn,
s->contents + my_offset);
bfd_put_32 (output_bfd, val | a2t2v5_func_addr_insn,
s->contents + my_offset + 4);
}
else
{
put_arm_insn (globals, output_bfd, (bfd_vma) a2t1_ldr_insn,
s->contents + my_offset);
put_arm_insn (globals, output_bfd, (bfd_vma) a2t2_bx_r12_insn,
s->contents + my_offset + 4);
bfd_put_32 (output_bfd, val | a2t3_func_addr_insn,
s->contents + my_offset + 8);
my_offset += 12;
}
}
BFD_ASSERT (my_offset <= globals->arm_glue_size);
return myh;
}
static int
elf32_arm_to_thumb_stub (struct bfd_link_info * info,
const char * name,
bfd * input_bfd,
bfd * output_bfd,
asection * input_section,
bfd_byte * hit_data,
asection * sym_sec,
bfd_vma offset,
bfd_signed_vma addend,
bfd_vma val,
char **error_message)
{
unsigned long int tmp;
bfd_vma my_offset;
asection * s;
long int ret_offset;
struct elf_link_hash_entry * myh;
struct elf32_arm_link_hash_table * globals;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name (globals->bfd_of_glue_owner,
ARM2THUMB_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
BFD_ASSERT (s->contents != NULL);
BFD_ASSERT (s->output_section != NULL);
myh = elf32_arm_create_thumb_stub (info, name, input_bfd, output_bfd,
sym_sec, val, s, error_message);
if (!myh)
return FALSE;
my_offset = myh->root.u.def.value;
tmp = bfd_get_32 (input_bfd, hit_data);
tmp = tmp & 0xFF000000;
ret_offset = (s->output_offset
+ my_offset
+ s->output_section->vma
- (input_section->output_offset
+ input_section->output_section->vma
+ offset + addend)
- 8);
tmp = tmp | ((ret_offset >> 2) & 0x00FFFFFF);
bfd_put_32 (output_bfd, (bfd_vma) tmp, hit_data - input_section->vma);
return TRUE;
}
static bfd_boolean
elf32_arm_to_thumb_export_stub (struct elf_link_hash_entry *h, void * inf)
{
struct bfd_link_info * info = (struct bfd_link_info *) inf;
asection * s;
struct elf_link_hash_entry * myh;
struct elf32_arm_link_hash_entry *eh;
struct elf32_arm_link_hash_table * globals;
asection *sec;
bfd_vma val;
char *error_message;
eh = elf32_arm_hash_entry (h);
if (eh->export_glue == NULL)
return TRUE;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name (globals->bfd_of_glue_owner,
ARM2THUMB_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
BFD_ASSERT (s->contents != NULL);
BFD_ASSERT (s->output_section != NULL);
sec = eh->export_glue->root.u.def.section;
BFD_ASSERT (sec->output_section != NULL);
val = eh->export_glue->root.u.def.value + sec->output_offset
+ sec->output_section->vma;
myh = elf32_arm_create_thumb_stub (info, h->root.root.string,
h->root.u.def.section->owner,
globals->obfd, sec, val, s,
&error_message);
BFD_ASSERT (myh);
return TRUE;
}
static bfd_vma
elf32_arm_bx_glue (struct bfd_link_info * info, int reg)
{
bfd_byte *p;
bfd_vma glue_addr;
asection *s;
struct elf32_arm_link_hash_table *globals;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (globals != NULL);
BFD_ASSERT (globals->bfd_of_glue_owner != NULL);
s = bfd_get_section_by_name (globals->bfd_of_glue_owner,
ARM_BX_GLUE_SECTION_NAME);
BFD_ASSERT (s != NULL);
BFD_ASSERT (s->contents != NULL);
BFD_ASSERT (s->output_section != NULL);
BFD_ASSERT (globals->bx_glue_offset[reg] & 2);
glue_addr = globals->bx_glue_offset[reg] & ~(bfd_vma)3;
if ((globals->bx_glue_offset[reg] & 1) == 0)
{
p = s->contents + glue_addr;
bfd_put_32 (globals->obfd, armbx1_tst_insn + (reg << 16), p);
bfd_put_32 (globals->obfd, armbx2_moveq_insn + reg, p + 4);
bfd_put_32 (globals->obfd, armbx3_bx_insn + reg, p + 8);
globals->bx_glue_offset[reg] |= 1;
}
return glue_addr + s->output_section->vma + s->output_offset;
}
static void
elf32_arm_begin_write_processing (bfd *abfd ATTRIBUTE_UNUSED,
struct bfd_link_info *link_info)
{
struct elf32_arm_link_hash_table * globals;
if (link_info == NULL)
return;
globals = elf32_arm_hash_table (link_info);
nothing to do. */
if (globals->use_blx)
return;
elf_link_hash_traverse (&globals->root, elf32_arm_to_thumb_export_stub,
link_info);
}
target. Return the real relocation. */
static int
arm_real_reloc_type (struct elf32_arm_link_hash_table * globals,
int r_type)
{
switch (r_type)
{
case R_ARM_TARGET1:
if (globals->target1_is_rel)
return R_ARM_REL32;
else
return R_ARM_ABS32;
case R_ARM_TARGET2:
return globals->target2_reloc;
default:
return r_type;
}
}
when resolving @dtpoff relocation.
This is PT_TLS segment p_vaddr. */
static bfd_vma
dtpoff_base (struct bfd_link_info *info)
{
if (elf_hash_table (info)->tls_sec == NULL)
return 0;
return elf_hash_table (info)->tls_sec->vma;
}
if STT_TLS virtual address is ADDRESS. */
static bfd_vma
tpoff (struct bfd_link_info *info, bfd_vma address)
{
struct elf_link_hash_table *htab = elf_hash_table (info);
bfd_vma base;
if (htab->tls_sec == NULL)
return 0;
base = align_power ((bfd_vma) TCB_SIZE, htab->tls_sec->alignment_power);
return address - htab->tls_sec->vma + base;
}
VALUE is the relocation value. */
static bfd_reloc_status_type
elf32_arm_abs12_reloc (bfd *abfd, void *data, bfd_vma value)
{
if (value > 0xfff)
return bfd_reloc_overflow;
value |= bfd_get_32 (abfd, data) & 0xfffff000;
bfd_put_32 (abfd, value, data);
return bfd_reloc_ok;
}
deal with group relocations. We return it in the form of an
encoded constant-and-rotation, together with the final residual. If n is
specified as less than zero, then final_residual is filled with the
input value and no further action is performed. */
static bfd_vma
calculate_group_reloc_mask (bfd_vma value, int n, bfd_vma *final_residual)
{
int current_n;
bfd_vma g_n;
bfd_vma encoded_g_n = 0;
bfd_vma residual = value;
for (current_n = 0; current_n <= n; current_n++)
{
int shift;
if (residual == 0)
shift = 0;
else
{
int msb;
align the resulting value to a 2-bit boundary. */
for (msb = 30; msb >= 0; msb -= 2)
if (residual & (3 << msb))
break;
is the greater. */
shift = msb - 6;
if (shift < 0)
shift = 0;
}
g_n = residual & (0xff << shift);
encoded_g_n = (g_n >> shift)
| ((g_n <= 0xff ? 0 : (32 - shift) / 2) << 8);
residual &= ~g_n;
}
*final_residual = residual;
return encoded_g_n;
}
Returns 1 if it is an ADD, -1 if it is a SUB, and 0 otherwise. */
static int
identify_add_or_sub (bfd_vma insn)
{
int opcode = insn & 0x1e00000;
if (opcode == 1 << 23)
return 1;
if (opcode == 1 << 22)
return -1;
return 0;
}
static bfd_reloc_status_type
elf32_arm_final_link_relocate (reloc_howto_type * howto,
bfd * input_bfd,
bfd * output_bfd,
asection * input_section,
bfd_byte * contents,
Elf_Internal_Rela * rel,
bfd_vma value,
struct bfd_link_info * info,
asection * sym_sec,
const char * sym_name,
int sym_flags,
struct elf_link_hash_entry * h,
bfd_boolean * unresolved_reloc_p,
char ** error_message)
{
unsigned long r_type = howto->type;
unsigned long r_symndx;
bfd_byte * hit_data = contents + rel->r_offset;
bfd * dynobj = NULL;
Elf_Internal_Shdr * symtab_hdr;
struct elf_link_hash_entry ** sym_hashes;
bfd_vma * local_got_offsets;
asection * sgot = NULL;
asection * splt = NULL;
asection * sreloc = NULL;
bfd_vma addend;
bfd_signed_vma signed_addend;
struct elf32_arm_link_hash_table * globals;
globals = elf32_arm_hash_table (info);
BFD_ASSERT (is_arm_elf (input_bfd));
target. We pick the right one here. */
r_type = arm_real_reloc_type (globals, r_type);
if (r_type != howto->type)
howto = elf32_arm_howto_from_type (r_type);
flag. Setting this more than once is redundant, but the cost is
not too high, and it keeps the code simple.
The test is done here, rather than somewhere else, because the
start address is only set just before the final link commences.
Note - if the user deliberately sets a start address of 0, the
flag will not be set. */
if (bfd_get_start_address (output_bfd) != 0)
elf_elfheader (output_bfd)->e_flags |= EF_ARM_HASENTRY;
dynobj = elf_hash_table (info)->dynobj;
if (dynobj)
{
sgot = bfd_get_section_by_name (dynobj, ".got");
splt = bfd_get_section_by_name (dynobj, ".plt");
}
symtab_hdr = & elf_symtab_hdr (input_bfd);
sym_hashes = elf_sym_hashes (input_bfd);
local_got_offsets = elf_local_got_offsets (input_bfd);
r_symndx = ELF32_R_SYM (rel->r_info);
if (globals->use_rel)
{
addend = bfd_get_32 (input_bfd, hit_data) & howto->src_mask;
if (addend & ((howto->src_mask + 1) >> 1))
{
signed_addend = -1;
signed_addend &= ~ howto->src_mask;
signed_addend |= addend;
}
else
signed_addend = addend;
}
else
addend = signed_addend = rel->r_addend;
switch (r_type)
{
case R_ARM_NONE:
marker. */
*unresolved_reloc_p = FALSE;
return bfd_reloc_ok;
case R_ARM_ABS12:
if (!globals->vxworks_p)
return elf32_arm_abs12_reloc (input_bfd, hit_data, value + addend);
case R_ARM_PC24:
case R_ARM_ABS32:
case R_ARM_ABS32_NOI:
case R_ARM_REL32:
case R_ARM_REL32_NOI:
case R_ARM_CALL:
case R_ARM_JUMP24:
case R_ARM_XPC25:
case R_ARM_PREL31:
case R_ARM_PLT32:
will use the symbol's value, which may point to a PLT entry, but we
don't need to handle that here. If we created a PLT entry, all
branches in this object should go to it. */
if ((r_type != R_ARM_ABS32 && r_type != R_ARM_REL32
&& r_type != R_ARM_ABS32_NOI && r_type != R_ARM_REL32_NOI)
&& h != NULL
&& splt != NULL
&& h->plt.offset != (bfd_vma) -1)
{
this function, it should not be known to bind locally. If
it were, we would have cleared the PLT entry. */
BFD_ASSERT (!SYMBOL_CALLS_LOCAL (info, h));
value = (splt->output_section->vma
+ splt->output_offset
+ h->plt.offset);
*unresolved_reloc_p = FALSE;
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
}
relocations are copied into the output file to be resolved at
run time. */
if ((info->shared || globals->root.is_relocatable_executable)
&& (input_section->flags & SEC_ALLOC)
&& !(elf32_arm_hash_table (info)->vxworks_p
&& strcmp (input_section->output_section->name,
".tls_vars") == 0)
&& ((r_type != R_ARM_REL32 && r_type != R_ARM_REL32_NOI)
|| !SYMBOL_CALLS_LOCAL (info, h))
&& (h == NULL
|| ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
|| h->root.type != bfd_link_hash_undefweak)
&& r_type != R_ARM_PC24
&& r_type != R_ARM_CALL
&& r_type != R_ARM_JUMP24
&& r_type != R_ARM_PREL31
&& r_type != R_ARM_PLT32)
{
Elf_Internal_Rela outrel;
bfd_byte *loc;
bfd_boolean skip, relocate;
*unresolved_reloc_p = FALSE;
if (sreloc == NULL)
{
const char * name;
name = (bfd_elf_string_from_elf_section
(input_bfd,
elf_elfheader (input_bfd)->e_shstrndx,
elf_section_data (input_section)->rel_hdr.sh_name));
if (name == NULL)
return bfd_reloc_notsupported;
BFD_ASSERT (reloc_section_p (globals, name, input_section));
sreloc = bfd_get_section_by_name (dynobj, name);
BFD_ASSERT (sreloc != NULL);
}
skip = FALSE;
relocate = FALSE;
outrel.r_addend = addend;
outrel.r_offset =
_bfd_elf_section_offset (output_bfd, info, input_section,
rel->r_offset);
if (outrel.r_offset == (bfd_vma) -1)
skip = TRUE;
else if (outrel.r_offset == (bfd_vma) -2)
skip = TRUE, relocate = TRUE;
outrel.r_offset += (input_section->output_section->vma
+ input_section->output_offset);
if (skip)
memset (&outrel, 0, sizeof outrel);
else if (h != NULL
&& h->dynindx != -1
&& (!info->shared
|| !info->symbolic
|| !h->def_regular))
outrel.r_info = ELF32_R_INFO (h->dynindx, r_type);
else
{
int symbol;
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
if (globals->symbian_p)
{
asection *osec;
can be relocated independently. Therefore, we
must indicate the segment to which this
relocation is relative. The BPABI allows us to
use any symbol in the right segment; we just use
the section symbol as it is convenient. (We
cannot use the symbol given by "h" directly as it
will not appear in the dynamic symbol table.)
Note that the dynamic linker ignores the section
symbol value, so we don't subtract osec->vma
from the emitted reloc addend. */
if (sym_sec)
osec = sym_sec->output_section;
else
osec = input_section->output_section;
symbol = elf_section_data (osec)->dynindx;
if (symbol == 0)
{
struct elf_link_hash_table *htab = elf_hash_table (info);
if ((osec->flags & SEC_READONLY) == 0
&& htab->data_index_section != NULL)
osec = htab->data_index_section;
else
osec = htab->text_index_section;
symbol = elf_section_data (osec)->dynindx;
}
BFD_ASSERT (symbol != 0);
}
else
relocate the text and data segments independently,
so the symbol does not matter. */
symbol = 0;
outrel.r_info = ELF32_R_INFO (symbol, R_ARM_RELATIVE);
if (globals->use_rel)
relocate = TRUE;
else
outrel.r_addend += value;
}
loc = sreloc->contents;
loc += sreloc->reloc_count++ * RELOC_SIZE (globals);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
fiddle with the addend. Otherwise, we need to include the symbol
value so that it becomes an addend for the dynamic reloc. */
if (! relocate)
return bfd_reloc_ok;
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
(bfd_vma) 0);
}
else switch (r_type)
{
case R_ARM_ABS12:
return elf32_arm_abs12_reloc (input_bfd, hit_data, value + addend);
case R_ARM_XPC25:
case R_ARM_CALL:
case R_ARM_JUMP24:
case R_ARM_PC24:
case R_ARM_PLT32:
{
bfd_vma from;
bfd_signed_vma branch_offset;
struct elf32_arm_stub_hash_entry *stub_entry = NULL;
from = (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
branch_offset = (bfd_signed_vma)(value - from);
if (r_type == R_ARM_XPC25)
{
instruction instead ? */
if (sym_flags != STT_ARM_TFUNC)
(*_bfd_error_handler)
(_("\%B: Warning: Arm BLX instruction targets Arm function '%s'."),
input_bfd,
h ? h->root.root.string : "(local)");
}
else if (r_type != R_ARM_CALL)
{
if (sym_flags == STT_ARM_TFUNC)
{
if (elf32_arm_to_thumb_stub (info, sym_name, input_bfd,
output_bfd, input_section,
hit_data, sym_sec, rel->r_offset,
signed_addend, value,
error_message))
return bfd_reloc_ok;
else
return bfd_reloc_dangerous;
}
}
destination is too far or we are changing mode. */
if (r_type == R_ARM_CALL)
{
if (branch_offset > ARM_MAX_FWD_BRANCH_OFFSET
|| branch_offset < ARM_MAX_BWD_BRANCH_OFFSET
|| sym_flags == STT_ARM_TFUNC)
{
branch to the local stub for this function. */
stub_entry = elf32_arm_get_stub_entry (input_section,
sym_sec, h,
rel, globals);
if (stub_entry != NULL)
value = (stub_entry->stub_offset
+ stub_entry->stub_sec->output_offset
+ stub_entry->stub_sec->output_section->vma);
}
}
where:
S is the address of the symbol in the relocation.
P is address of the instruction being relocated.
A is the addend (extracted from the instruction) in bytes.
S is held in 'value'.
P is the base address of the section containing the
instruction plus the offset of the reloc into that
section, ie:
(input_section->output_section->vma +
input_section->output_offset +
rel->r_offset).
A is the addend, converted into bytes, ie:
(signed_addend * 4)
Note: None of these operations have knowledge of the pipeline
size of the processor, thus it is up to the assembler to
encode this information into the addend. */
value -= (input_section->output_section->vma
+ input_section->output_offset);
value -= rel->r_offset;
if (globals->use_rel)
value += (signed_addend << howto->size);
else
value += signed_addend;
signed_addend = value;
signed_addend >>= howto->rightshift;
the next instruction. */
if (h && h->root.type == bfd_link_hash_undefweak)
{
value = (bfd_get_32 (input_bfd, hit_data) & 0xf0000000)
| 0x0affffff;
}
else
{
if ( signed_addend > ((bfd_signed_vma) (howto->dst_mask >> 1))
|| signed_addend < - ((bfd_signed_vma) ((howto->dst_mask + 1) >> 1)))
return bfd_reloc_overflow;
addend = (value & 2);
value = (signed_addend & howto->dst_mask)
| (bfd_get_32 (input_bfd, hit_data) & (~ howto->dst_mask));
if (sym_flags == STT_ARM_TFUNC)
{
if (addend)
value |= (1 << 24);
else
value &= ~(bfd_vma)(1 << 24);
}
if (r_type == R_ARM_CALL)
{
case, mode switching is performed by the stub. */
if (sym_flags == STT_ARM_TFUNC && !stub_entry)
value |= (1 << 28);
else
{
value &= ~(bfd_vma)(1 << 28);
value |= (1 << 24);
}
}
}
}
break;
case R_ARM_ABS32:
value += addend;
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
break;
case R_ARM_ABS32_NOI:
value += addend;
break;
case R_ARM_REL32:
value += addend;
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
value -= (input_section->output_section->vma
+ input_section->output_offset + rel->r_offset);
break;
case R_ARM_REL32_NOI:
value += addend;
value -= (input_section->output_section->vma
+ input_section->output_offset + rel->r_offset);
break;
case R_ARM_PREL31:
value -= (input_section->output_section->vma
+ input_section->output_offset + rel->r_offset);
value += signed_addend;
if (! h || h->root.type != bfd_link_hash_undefweak)
{
if ((value ^ (value >> 1)) & (1 << 30))
return bfd_reloc_overflow;
}
value &= 0x7fffffff;
value |= (bfd_get_32 (input_bfd, hit_data) & 0x80000000);
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
break;
}
bfd_put_32 (input_bfd, value, hit_data);
return bfd_reloc_ok;
case R_ARM_ABS8:
value += addend;
if ((long) value > 0x7f || (long) value < -0x80)
return bfd_reloc_overflow;
bfd_put_8 (input_bfd, value, hit_data);
return bfd_reloc_ok;
case R_ARM_ABS16:
value += addend;
if ((long) value > 0x7fff || (long) value < -0x8000)
return bfd_reloc_overflow;
bfd_put_16 (input_bfd, value, hit_data);
return bfd_reloc_ok;
case R_ARM_THM_ABS5:
if (globals->use_rel)
{
addend = bfd_get_16 (input_bfd, hit_data) & howto->src_mask;
addend >>= howto->rightshift;
}
value += addend;
if ((long) value > 0x1f || (long) value < -0x10)
return bfd_reloc_overflow;
value |= bfd_get_16 (input_bfd, hit_data) & 0xf83f;
bfd_put_16 (input_bfd, value, hit_data);
return bfd_reloc_ok;
case R_ARM_THM_ALU_PREL_11_0:
{
bfd_vma insn;
bfd_signed_vma relocation;
insn = (bfd_get_16 (input_bfd, hit_data) << 16)
| bfd_get_16 (input_bfd, hit_data + 2);
if (globals->use_rel)
{
signed_addend = (insn & 0xff) | ((insn & 0x7000) >> 4)
| ((insn & (1 << 26)) >> 15);
if (insn & 0xf00000)
signed_addend = -signed_addend;
}
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
value = abs (relocation);
if (value >= 0x1000)
return bfd_reloc_overflow;
insn = (insn & 0xfb0f8f00) | (value & 0xff)
| ((value & 0x700) << 4)
| ((value & 0x800) << 15);
if (relocation < 0)
insn |= 0xa00000;
bfd_put_16 (input_bfd, insn >> 16, hit_data);
bfd_put_16 (input_bfd, insn & 0xffff, hit_data + 2);
return bfd_reloc_ok;
}
case R_ARM_THM_PC12:
{
bfd_vma insn;
bfd_signed_vma relocation;
insn = (bfd_get_16 (input_bfd, hit_data) << 16)
| bfd_get_16 (input_bfd, hit_data + 2);
if (globals->use_rel)
{
signed_addend = insn & 0xfff;
if (!(insn & (1 << 23)))
signed_addend = -signed_addend;
}
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
value = abs (relocation);
if (value >= 0x1000)
return bfd_reloc_overflow;
insn = (insn & 0xff7ff000) | value;
if (relocation >= 0)
insn |= (1 << 23);
bfd_put_16 (input_bfd, insn >> 16, hit_data);
bfd_put_16 (input_bfd, insn & 0xffff, hit_data + 2);
return bfd_reloc_ok;
}
case R_ARM_THM_XPC22:
case R_ARM_THM_CALL:
case R_ARM_THM_JUMP24:
{
bfd_vma relocation;
bfd_vma reloc_sign;
bfd_boolean overflow = FALSE;
bfd_vma upper_insn = bfd_get_16 (input_bfd, hit_data);
bfd_vma lower_insn = bfd_get_16 (input_bfd, hit_data + 2);
bfd_signed_vma reloc_signed_max;
bfd_signed_vma reloc_signed_min;
bfd_vma check;
bfd_signed_vma signed_check;
int bitsize;
int thumb2 = using_thumb2 (globals);
the next instruction. */
if (h && h->root.type == bfd_link_hash_undefweak)
{
bfd_put_16 (input_bfd, 0xe000, hit_data);
bfd_put_16 (input_bfd, 0xbf00, hit_data + 2);
return bfd_reloc_ok;
}
with Thumb-1) involving the J1 and J2 bits. */
if (globals->use_rel)
{
bfd_vma s = (upper_insn & (1 << 10)) >> 10;
bfd_vma upper = upper_insn & 0x3ff;
bfd_vma lower = lower_insn & 0x7ff;
bfd_vma j1 = (lower_insn & (1 << 13)) >> 13;
bfd_vma j2 = (lower_insn & (1 << 11)) >> 11;
bfd_vma i1 = j1 ^ s ? 0 : 1;
bfd_vma i2 = j2 ^ s ? 0 : 1;
addend = (i1 << 23) | (i2 << 22) | (upper << 12) | (lower << 1);
addend = (addend | ((s ? 0 : 1) << 24)) - (1 << 24);
signed_addend = addend;
}
if (r_type == R_ARM_THM_XPC22)
{
instruction instead ? */
if (sym_flags == STT_ARM_TFUNC)
(*_bfd_error_handler)
(_("%B: Warning: Thumb BLX instruction targets thumb function '%s'."),
input_bfd,
h ? h->root.root.string : "(local)");
}
else
{
If it is a call relative to a section name, then it is not a
function call at all, but rather a long jump. Calls through
the PLT do not require stubs. */
if (sym_flags != STT_ARM_TFUNC && sym_flags != STT_SECTION
&& (h == NULL || splt == NULL
|| h->plt.offset == (bfd_vma) -1))
{
if (globals->use_blx && r_type == R_ARM_THM_CALL)
{
lower_insn = (lower_insn & ~0x1000) | 0x0800;
}
else if (r_type != R_ARM_THM_CALL)
{
if (elf32_thumb_to_arm_stub
(info, sym_name, input_bfd, output_bfd, input_section,
hit_data, sym_sec, rel->r_offset, signed_addend, value,
error_message))
return bfd_reloc_ok;
else
return bfd_reloc_dangerous;
}
}
else if (sym_flags == STT_ARM_TFUNC && globals->use_blx
&& r_type == R_ARM_THM_CALL)
{
lower_insn |= 0x1800;
}
}
if (h != NULL && splt != NULL && h->plt.offset != (bfd_vma) -1)
{
value = (splt->output_section->vma
+ splt->output_offset
+ h->plt.offset);
if (globals->use_blx && r_type == R_ARM_THM_CALL)
{
BL to a BLX instruction to call the ARM-mode PLT entry. */
lower_insn = (lower_insn & ~0x1000) | 0x0800;
}
else
value -= PLT_THUMB_STUB_SIZE;
*unresolved_reloc_p = FALSE;
}
if (r_type == R_ARM_THM_CALL)
{
is too far. */
bfd_vma from;
bfd_signed_vma branch_offset;
struct elf32_arm_stub_hash_entry *stub_entry = NULL;
from = (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
branch_offset = (bfd_signed_vma)(value - from);
if ((!thumb2
&& (branch_offset > THM_MAX_FWD_BRANCH_OFFSET
|| (branch_offset < THM_MAX_BWD_BRANCH_OFFSET)))
||
(thumb2
&& (branch_offset > THM2_MAX_FWD_BRANCH_OFFSET
|| (branch_offset < THM2_MAX_BWD_BRANCH_OFFSET)))
|| ((sym_flags != STT_ARM_TFUNC) && !globals->use_blx))
{
redirect the branch to the local stub for this
function. */
stub_entry = elf32_arm_get_stub_entry (input_section,
sym_sec, h,
rel, globals);
if (stub_entry != NULL)
value = (stub_entry->stub_offset
+ stub_entry->stub_sec->output_offset
+ stub_entry->stub_sec->output_section->vma);
if (globals->use_blx)
{
if ((stub_entry
&& !arm_stub_is_thumb (stub_entry->stub_type))
|| (sym_flags != STT_ARM_TFUNC))
lower_insn = (lower_insn & ~0x1000) | 0x0800;
}
}
}
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
check = relocation >> howto->rightshift;
leading 1 bits (assuming twos complement). */
if ((bfd_signed_vma) relocation >= 0)
signed_check = check;
else
signed_check = check | ~((bfd_vma) -1 >> howto->rightshift);
this relocation according to whether we're relocating for
Thumb-2 or not. */
bitsize = howto->bitsize;
if (!thumb2)
bitsize -= 2;
reloc_signed_max = ((1 << (bitsize - 1)) - 1) >> howto->rightshift;
reloc_signed_min = ~reloc_signed_max;
if (signed_check > reloc_signed_max || signed_check < reloc_signed_min)
overflow = TRUE;
if ((lower_insn & 0x5000) == 0x4000)
to a word boundary. This follows the semantics of the instruction
which specifies that bit 1 of the target address will come from bit
1 of the base address. */
relocation = (relocation + 2) & ~ 3;
We use the Thumb-2 encoding, which is safe even if dealing with
a Thumb-1 instruction by virtue of our overflow check above. */
reloc_sign = (signed_check < 0) ? 1 : 0;
upper_insn = (upper_insn & ~(bfd_vma) 0x7ff)
| ((relocation >> 12) & 0x3ff)
| (reloc_sign << 10);
lower_insn = (lower_insn & ~(bfd_vma) 0x2fff)
| (((!((relocation >> 23) & 1)) ^ reloc_sign) << 13)
| (((!((relocation >> 22) & 1)) ^ reloc_sign) << 11)
| ((relocation >> 1) & 0x7ff);
bfd_put_16 (input_bfd, upper_insn, hit_data);
bfd_put_16 (input_bfd, lower_insn, hit_data + 2);
return (overflow ? bfd_reloc_overflow : bfd_reloc_ok);
}
break;
case R_ARM_THM_JUMP19:
{
bfd_vma relocation;
bfd_boolean overflow = FALSE;
bfd_vma upper_insn = bfd_get_16 (input_bfd, hit_data);
bfd_vma lower_insn = bfd_get_16 (input_bfd, hit_data + 2);
bfd_signed_vma reloc_signed_max = 0xffffe;
bfd_signed_vma reloc_signed_min = -0x100000;
bfd_signed_vma signed_check;
and squish the two 11 bit pieces together. */
if (globals->use_rel)
{
bfd_vma S = (upper_insn & 0x0400) >> 10;
bfd_vma upper = (upper_insn & 0x003f);
bfd_vma J1 = (lower_insn & 0x2000) >> 13;
bfd_vma J2 = (lower_insn & 0x0800) >> 11;
bfd_vma lower = (lower_insn & 0x07ff);
upper |= J1 << 6;
upper |= J2 << 7;
upper |= (!S) << 8;
upper -= 0x0100;
addend = (upper << 12) | (lower << 1);
signed_addend = addend;
}
if (h != NULL && splt != NULL && h->plt.offset != (bfd_vma) -1)
{
value = (splt->output_section->vma
+ splt->output_offset
+ h->plt.offset);
value -= PLT_THUMB_STUB_SIZE;
*unresolved_reloc_p = FALSE;
}
use this for tail calls. */
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
signed_check = (bfd_signed_vma) relocation;
if (signed_check > reloc_signed_max || signed_check < reloc_signed_min)
overflow = TRUE;
{
bfd_vma S = (relocation & 0x00100000) >> 20;
bfd_vma J2 = (relocation & 0x00080000) >> 19;
bfd_vma J1 = (relocation & 0x00040000) >> 18;
bfd_vma hi = (relocation & 0x0003f000) >> 12;
bfd_vma lo = (relocation & 0x00000ffe) >> 1;
upper_insn = (upper_insn & 0xfbc0) | (S << 10) | hi;
lower_insn = (lower_insn & 0xd000) | (J1 << 13) | (J2 << 11) | lo;
}
bfd_put_16 (input_bfd, upper_insn, hit_data);
bfd_put_16 (input_bfd, lower_insn, hit_data + 2);
return (overflow ? bfd_reloc_overflow : bfd_reloc_ok);
}
case R_ARM_THM_JUMP11:
case R_ARM_THM_JUMP8:
case R_ARM_THM_JUMP6:
{
bfd_signed_vma relocation;
bfd_signed_vma reloc_signed_max = (1 << (howto->bitsize - 1)) - 1;
bfd_signed_vma reloc_signed_min = ~ reloc_signed_max;
bfd_signed_vma signed_check;
if (r_type == R_ARM_THM_JUMP6)
reloc_signed_min = 0;
if (globals->use_rel)
{
addend = bfd_get_16 (input_bfd, hit_data) & howto->src_mask;
if (addend & ((howto->src_mask + 1) >> 1))
{
signed_addend = -1;
signed_addend &= ~ howto->src_mask;
signed_addend |= addend;
}
else
signed_addend = addend;
undo this, so that we can perform the address calculation
in terms of bytes. */
signed_addend <<= howto->rightshift;
}
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
relocation >>= howto->rightshift;
signed_check = relocation;
if (r_type == R_ARM_THM_JUMP6)
relocation = ((relocation & 0x0020) << 4) | ((relocation & 0x001f) << 3);
else
relocation &= howto->dst_mask;
relocation |= (bfd_get_16 (input_bfd, hit_data) & (~ howto->dst_mask));
bfd_put_16 (input_bfd, relocation, hit_data);
if (signed_check > reloc_signed_max || signed_check < reloc_signed_min)
return bfd_reloc_overflow;
return bfd_reloc_ok;
}
case R_ARM_ALU_PCREL7_0:
case R_ARM_ALU_PCREL15_8:
case R_ARM_ALU_PCREL23_15:
{
bfd_vma insn;
bfd_vma relocation;
insn = bfd_get_32 (input_bfd, hit_data);
if (globals->use_rel)
{
addend = (insn & 0xff) << ((insn & 0xf00) >> 7);
signed_addend = addend;
}
relocation = value + signed_addend;
relocation -= (input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset);
insn = (insn & ~0xfff)
| ((howto->bitpos << 7) & 0xf00)
| ((relocation >> howto->bitpos) & 0xff);
bfd_put_32 (input_bfd, value, hit_data);
}
return bfd_reloc_ok;
case R_ARM_GNU_VTINHERIT:
case R_ARM_GNU_VTENTRY:
return bfd_reloc_ok;
case R_ARM_GOTOFF32:
global offset table. */
BFD_ASSERT (sgot != NULL);
if (sgot == NULL)
return bfd_reloc_notsupported;
address by one, so that attempts to call the function pointer will
correctly interpret it as Thumb code. */
if (sym_flags == STT_ARM_TFUNC)
value += 1;
calculation. We always want the start of .got. If we
define _GLOBAL_OFFSET_TABLE in a different way, as is
permitted by the ABI, we might have to change this
calculation. */
value -= sgot->output_section->vma;
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
case R_ARM_GOTPC:
BFD_ASSERT (sgot != NULL);
if (sgot == NULL)
return bfd_reloc_notsupported;
*unresolved_reloc_p = FALSE;
value = sgot->output_section->vma;
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
case R_ARM_GOT32:
case R_ARM_GOT_PREL:
global offset table. */
if (sgot == NULL)
return bfd_reloc_notsupported;
if (h != NULL)
{
bfd_vma off;
bfd_boolean dyn;
off = h->got.offset;
BFD_ASSERT (off != (bfd_vma) -1);
dyn = globals->root.dynamic_sections_created;
if (! WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, info->shared, h)
|| (info->shared
&& SYMBOL_REFERENCES_LOCAL (info, h))
|| (ELF_ST_VISIBILITY (h->other)
&& h->root.type == bfd_link_hash_undefweak))
{
and the symbol is defined locally. We must initialize this
entry in the global offset table. Since the offset must
always be a multiple of 4, we use the least significant bit
to record whether we have initialized it already.
When doing a dynamic link, we create a .rel(a).got relocation
entry to initialize the value. This is done in the
finish_dynamic_symbol routine. */
if ((off & 1) != 0)
off &= ~1;
else
{
adjust the address by one, so that attempts to
call the function pointer will correctly
interpret it as Thumb code. */
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
bfd_put_32 (output_bfd, value, sgot->contents + off);
h->got.offset |= 1;
}
}
else
*unresolved_reloc_p = FALSE;
value = sgot->output_offset + off;
}
else
{
bfd_vma off;
BFD_ASSERT (local_got_offsets != NULL &&
local_got_offsets[r_symndx] != (bfd_vma) -1);
off = local_got_offsets[r_symndx];
least significant bit to record whether we have already
generated the necessary reloc. */
if ((off & 1) != 0)
off &= ~1;
else
{
adjust the address by one, so that attempts to
call the function pointer will correctly
interpret it as Thumb code. */
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
if (globals->use_rel)
bfd_put_32 (output_bfd, value, sgot->contents + off);
if (info->shared)
{
asection * srelgot;
Elf_Internal_Rela outrel;
bfd_byte *loc;
srelgot = (bfd_get_section_by_name
(dynobj, RELOC_SECTION (globals, ".got")));
BFD_ASSERT (srelgot != NULL);
outrel.r_addend = addend + value;
outrel.r_offset = (sgot->output_section->vma
+ sgot->output_offset
+ off);
outrel.r_info = ELF32_R_INFO (0, R_ARM_RELATIVE);
loc = srelgot->contents;
loc += srelgot->reloc_count++ * RELOC_SIZE (globals);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
}
local_got_offsets[r_symndx] |= 1;
}
value = sgot->output_offset + off;
}
if (r_type != R_ARM_GOT32)
value += sgot->output_section->vma;
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
case R_ARM_TLS_LDO32:
value = value - dtpoff_base (info);
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
case R_ARM_TLS_LDM32:
{
bfd_vma off;
if (globals->sgot == NULL)
abort ();
off = globals->tls_ldm_got.offset;
if ((off & 1) != 0)
off &= ~1;
else
{
for it. */
if (info->shared)
{
Elf_Internal_Rela outrel;
bfd_byte *loc;
if (globals->srelgot == NULL)
abort ();
outrel.r_addend = 0;
outrel.r_offset = (globals->sgot->output_section->vma
+ globals->sgot->output_offset + off);
outrel.r_info = ELF32_R_INFO (0, R_ARM_TLS_DTPMOD32);
if (globals->use_rel)
bfd_put_32 (output_bfd, outrel.r_addend,
globals->sgot->contents + off);
loc = globals->srelgot->contents;
loc += globals->srelgot->reloc_count++ * RELOC_SIZE (globals);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
}
else
bfd_put_32 (output_bfd, 1, globals->sgot->contents + off);
globals->tls_ldm_got.offset |= 1;
}
value = globals->sgot->output_section->vma + globals->sgot->output_offset + off
- (input_section->output_section->vma + input_section->output_offset + rel->r_offset);
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
}
case R_ARM_TLS_GD32:
case R_ARM_TLS_IE32:
{
bfd_vma off;
int indx;
char tls_type;
if (globals->sgot == NULL)
abort ();
indx = 0;
if (h != NULL)
{
bfd_boolean dyn;
dyn = globals->root.dynamic_sections_created;
if (WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, info->shared, h)
&& (!info->shared
|| !SYMBOL_REFERENCES_LOCAL (info, h)))
{
*unresolved_reloc_p = FALSE;
indx = h->dynindx;
}
off = h->got.offset;
tls_type = ((struct elf32_arm_link_hash_entry *) h)->tls_type;
}
else
{
if (local_got_offsets == NULL)
abort ();
off = local_got_offsets[r_symndx];
tls_type = elf32_arm_local_got_tls_type (input_bfd)[r_symndx];
}
if (tls_type == GOT_UNKNOWN)
abort ();
if ((off & 1) != 0)
off &= ~1;
else
{
bfd_boolean need_relocs = FALSE;
Elf_Internal_Rela outrel;
bfd_byte *loc = NULL;
int cur_off = off;
now, and emit any relocations. If both an IE GOT and a
GD GOT are necessary, we emit the GD first. */
if ((info->shared || indx != 0)
&& (h == NULL
|| ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
|| h->root.type != bfd_link_hash_undefweak))
{
need_relocs = TRUE;
if (globals->srelgot == NULL)
abort ();
loc = globals->srelgot->contents;
loc += globals->srelgot->reloc_count * RELOC_SIZE (globals);
}
if (tls_type & GOT_TLS_GD)
{
if (need_relocs)
{
outrel.r_addend = 0;
outrel.r_offset = (globals->sgot->output_section->vma
+ globals->sgot->output_offset
+ cur_off);
outrel.r_info = ELF32_R_INFO (indx, R_ARM_TLS_DTPMOD32);
if (globals->use_rel)
bfd_put_32 (output_bfd, outrel.r_addend,
globals->sgot->contents + cur_off);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
globals->srelgot->reloc_count++;
loc += RELOC_SIZE (globals);
if (indx == 0)
bfd_put_32 (output_bfd, value - dtpoff_base (info),
globals->sgot->contents + cur_off + 4);
else
{
outrel.r_addend = 0;
outrel.r_info = ELF32_R_INFO (indx,
R_ARM_TLS_DTPOFF32);
outrel.r_offset += 4;
if (globals->use_rel)
bfd_put_32 (output_bfd, outrel.r_addend,
globals->sgot->contents + cur_off + 4);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
globals->srelgot->reloc_count++;
loc += RELOC_SIZE (globals);
}
}
else
{
general dynamic reference, then we must be in a
static link or an executable link with the
symbol binding locally. Mark it as belonging
to module 1, the executable. */
bfd_put_32 (output_bfd, 1,
globals->sgot->contents + cur_off);
bfd_put_32 (output_bfd, value - dtpoff_base (info),
globals->sgot->contents + cur_off + 4);
}
cur_off += 8;
}
if (tls_type & GOT_TLS_IE)
{
if (need_relocs)
{
if (indx == 0)
outrel.r_addend = value - dtpoff_base (info);
else
outrel.r_addend = 0;
outrel.r_offset = (globals->sgot->output_section->vma
+ globals->sgot->output_offset
+ cur_off);
outrel.r_info = ELF32_R_INFO (indx, R_ARM_TLS_TPOFF32);
if (globals->use_rel)
bfd_put_32 (output_bfd, outrel.r_addend,
globals->sgot->contents + cur_off);
SWAP_RELOC_OUT (globals) (output_bfd, &outrel, loc);
globals->srelgot->reloc_count++;
loc += RELOC_SIZE (globals);
}
else
bfd_put_32 (output_bfd, tpoff (info, value),
globals->sgot->contents + cur_off);
cur_off += 4;
}
if (h != NULL)
h->got.offset |= 1;
else
local_got_offsets[r_symndx] |= 1;
}
if ((tls_type & GOT_TLS_GD) && r_type != R_ARM_TLS_GD32)
off += 8;
value = globals->sgot->output_section->vma + globals->sgot->output_offset + off
- (input_section->output_section->vma + input_section->output_offset + rel->r_offset);
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
}
case R_ARM_TLS_LE32:
if (info->shared)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): R_ARM_TLS_LE32 relocation not permitted in shared object"),
input_bfd, input_section,
(long) rel->r_offset, howto->name);
return FALSE;
}
else
value = tpoff (info, value);
return _bfd_final_link_relocate (howto, input_bfd, input_section,
contents, rel->r_offset, value,
rel->r_addend);
case R_ARM_V4BX:
if (globals->fix_v4bx)
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
BFD_ASSERT ((insn & 0x0ffffff0) == 0x012fff10);
if (globals->fix_v4bx == 2 && (insn & 0xf) != 0xf)
{
bfd_vma glue_addr;
glue_addr = elf32_arm_bx_glue (info, insn & 0xf);
glue_addr -= input_section->output_section->vma
+ input_section->output_offset
+ rel->r_offset + 8;
insn = (insn & 0xf0000000) | 0x0a000000
| ((glue_addr >> 2) & 0x00ffffff);
}
else
{
(highest four bits). Other bits encode MOV PC,Rm. */
insn = (insn & 0xf000000f) | 0x01a0f000;
}
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
case R_ARM_MOVW_ABS_NC:
case R_ARM_MOVT_ABS:
case R_ARM_MOVW_PREL_NC:
case R_ARM_MOVT_PREL:
we assume the segment base to be zero, as for the group relocations.
Thus R_ARM_MOVW_BREL_NC has the same semantics as R_ARM_MOVW_ABS_NC
and R_ARM_MOVT_BREL has the same semantics as R_ARM_MOVT_ABS. */
case R_ARM_MOVW_BREL_NC:
case R_ARM_MOVW_BREL:
case R_ARM_MOVT_BREL:
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
if (globals->use_rel)
{
addend = ((insn >> 4) & 0xf000) | (insn & 0xfff);
signed_addend = (addend ^ 0x8000) - 0x8000;
}
value += signed_addend;
if (r_type == R_ARM_MOVW_PREL_NC || r_type == R_ARM_MOVT_PREL)
value -= (input_section->output_section->vma
+ input_section->output_offset + rel->r_offset);
if (r_type == R_ARM_MOVW_BREL && value >= 0x10000)
return bfd_reloc_overflow;
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
if (r_type == R_ARM_MOVT_ABS || r_type == R_ARM_MOVT_PREL
|| r_type == R_ARM_MOVT_BREL)
value >>= 16;
insn &= 0xfff0f000;
insn |= value & 0xfff;
insn |= (value & 0xf000) << 4;
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
case R_ARM_THM_MOVW_ABS_NC:
case R_ARM_THM_MOVT_ABS:
case R_ARM_THM_MOVW_PREL_NC:
case R_ARM_THM_MOVT_PREL:
we assume the segment base to be zero, as for the above relocations.
Thus R_ARM_THM_MOVW_BREL_NC has the same semantics as
R_ARM_THM_MOVW_ABS_NC and R_ARM_THM_MOVT_BREL has the same semantics
as R_ARM_THM_MOVT_ABS. */
case R_ARM_THM_MOVW_BREL_NC:
case R_ARM_THM_MOVW_BREL:
case R_ARM_THM_MOVT_BREL:
{
bfd_vma insn;
insn = bfd_get_16 (input_bfd, hit_data) << 16;
insn |= bfd_get_16 (input_bfd, hit_data + 2);
if (globals->use_rel)
{
addend = ((insn >> 4) & 0xf000)
| ((insn >> 15) & 0x0800)
| ((insn >> 4) & 0x0700)
| (insn & 0x00ff);
signed_addend = (addend ^ 0x8000) - 0x8000;
}
value += signed_addend;
if (r_type == R_ARM_THM_MOVW_PREL_NC || r_type == R_ARM_THM_MOVT_PREL)
value -= (input_section->output_section->vma
+ input_section->output_offset + rel->r_offset);
if (r_type == R_ARM_THM_MOVW_BREL && value >= 0x10000)
return bfd_reloc_overflow;
if (sym_flags == STT_ARM_TFUNC)
value |= 1;
if (r_type == R_ARM_THM_MOVT_ABS || r_type == R_ARM_THM_MOVT_PREL
|| r_type == R_ARM_THM_MOVT_BREL)
value >>= 16;
insn &= 0xfbf08f00;
insn |= (value & 0xf000) << 4;
insn |= (value & 0x0800) << 15;
insn |= (value & 0x0700) << 4;
insn |= (value & 0x00ff);
bfd_put_16 (input_bfd, insn >> 16, hit_data);
bfd_put_16 (input_bfd, insn & 0xffff, hit_data + 2);
}
return bfd_reloc_ok;
case R_ARM_ALU_PC_G0_NC:
case R_ARM_ALU_PC_G1_NC:
case R_ARM_ALU_PC_G0:
case R_ARM_ALU_PC_G1:
case R_ARM_ALU_PC_G2:
case R_ARM_ALU_SB_G0_NC:
case R_ARM_ALU_SB_G1_NC:
case R_ARM_ALU_SB_G0:
case R_ARM_ALU_SB_G1:
case R_ARM_ALU_SB_G2:
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
bfd_vma pc = input_section->output_section->vma
+ input_section->output_offset + rel->r_offset;
It is not clear how to obtain this OS-dependent value, so we
make an arbitrary choice of zero. */
bfd_vma sb = 0;
bfd_vma residual;
bfd_vma g_n;
bfd_signed_vma signed_value;
int group = 0;
switch (r_type)
{
case R_ARM_ALU_PC_G0_NC:
case R_ARM_ALU_PC_G0:
case R_ARM_ALU_SB_G0_NC:
case R_ARM_ALU_SB_G0:
group = 0;
break;
case R_ARM_ALU_PC_G1_NC:
case R_ARM_ALU_PC_G1:
case R_ARM_ALU_SB_G1_NC:
case R_ARM_ALU_SB_G1:
group = 1;
break;
case R_ARM_ALU_PC_G2:
case R_ARM_ALU_SB_G2:
group = 2;
break;
default:
abort ();
}
have already been fetched for us. */
if (globals->use_rel)
{
int negative;
bfd_vma constant = insn & 0xff;
bfd_vma rotation = (insn & 0xf00) >> 8;
if (rotation == 0)
signed_addend = constant;
else
{
rotation is stored in multiples of 2 bits. */
rotation *= 2;
signed_addend = (constant >> rotation) |
(constant << (8 * sizeof (bfd_vma) - rotation));
}
(For REL, this determines the sign of the addend.) */
negative = identify_add_or_sub (insn);
if (negative == 0)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): Only ADD or SUB instructions are allowed for ALU group relocations"),
input_bfd, input_section,
(long) rel->r_offset, howto->name);
return bfd_reloc_overflow;
}
signed_addend *= negative;
}
if (r_type == R_ARM_ALU_PC_G0_NC
|| r_type == R_ARM_ALU_PC_G1_NC
|| r_type == R_ARM_ALU_PC_G0
|| r_type == R_ARM_ALU_PC_G1
|| r_type == R_ARM_ALU_PC_G2)
signed_value = value - pc + signed_addend;
else
signed_value = value - sb + signed_addend;
Thumb bit in the address. */
if (sym_flags == STT_ARM_TFUNC)
signed_value |= 1;
constant-with-rotation format. */
g_n = calculate_group_reloc_mask (abs (signed_value), group,
&residual);
if ((r_type == R_ARM_ALU_PC_G0
|| r_type == R_ARM_ALU_PC_G1
|| r_type == R_ARM_ALU_PC_G2
|| r_type == R_ARM_ALU_SB_G0
|| r_type == R_ARM_ALU_SB_G1
|| r_type == R_ARM_ALU_SB_G2) && residual != 0)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): Overflow whilst splitting 0x%lx for group relocation %s"),
input_bfd, input_section,
(long) rel->r_offset, abs (signed_value), howto->name);
return bfd_reloc_overflow;
}
not to destroy the S bit. */
insn &= 0xff1ff000;
place is negative. */
if (signed_value < 0)
insn |= 1 << 22;
else
insn |= 1 << 23;
insn |= g_n;
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
case R_ARM_LDR_PC_G0:
case R_ARM_LDR_PC_G1:
case R_ARM_LDR_PC_G2:
case R_ARM_LDR_SB_G0:
case R_ARM_LDR_SB_G1:
case R_ARM_LDR_SB_G2:
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
bfd_vma pc = input_section->output_section->vma
+ input_section->output_offset + rel->r_offset;
bfd_vma sb = 0;
bfd_vma residual;
bfd_signed_vma signed_value;
int group = 0;
switch (r_type)
{
case R_ARM_LDR_PC_G0:
case R_ARM_LDR_SB_G0:
group = 0;
break;
case R_ARM_LDR_PC_G1:
case R_ARM_LDR_SB_G1:
group = 1;
break;
case R_ARM_LDR_PC_G2:
case R_ARM_LDR_SB_G2:
group = 2;
break;
default:
abort ();
}
have already been fetched for us. */
if (globals->use_rel)
{
int negative = (insn & (1 << 23)) ? 1 : -1;
signed_addend = negative * (insn & 0xfff);
}
if (r_type == R_ARM_LDR_PC_G0
|| r_type == R_ARM_LDR_PC_G1
|| r_type == R_ARM_LDR_PC_G2)
signed_value = value - pc + signed_addend;
else
signed_value = value - sb + signed_addend;
the residual at that stage. */
calculate_group_reloc_mask (abs (signed_value), group - 1, &residual);
if (residual >= 0x1000)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): Overflow whilst splitting 0x%lx for group relocation %s"),
input_bfd, input_section,
(long) rel->r_offset, abs (signed_value), howto->name);
return bfd_reloc_overflow;
}
insn &= 0xff7ff000;
if (signed_value >= 0)
insn |= 1 << 23;
insn |= residual;
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
case R_ARM_LDRS_PC_G0:
case R_ARM_LDRS_PC_G1:
case R_ARM_LDRS_PC_G2:
case R_ARM_LDRS_SB_G0:
case R_ARM_LDRS_SB_G1:
case R_ARM_LDRS_SB_G2:
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
bfd_vma pc = input_section->output_section->vma
+ input_section->output_offset + rel->r_offset;
bfd_vma sb = 0;
bfd_vma residual;
bfd_signed_vma signed_value;
int group = 0;
switch (r_type)
{
case R_ARM_LDRS_PC_G0:
case R_ARM_LDRS_SB_G0:
group = 0;
break;
case R_ARM_LDRS_PC_G1:
case R_ARM_LDRS_SB_G1:
group = 1;
break;
case R_ARM_LDRS_PC_G2:
case R_ARM_LDRS_SB_G2:
group = 2;
break;
default:
abort ();
}
have already been fetched for us. */
if (globals->use_rel)
{
int negative = (insn & (1 << 23)) ? 1 : -1;
signed_addend = negative * (((insn & 0xf00) >> 4) + (insn & 0xf));
}
if (r_type == R_ARM_LDRS_PC_G0
|| r_type == R_ARM_LDRS_PC_G1
|| r_type == R_ARM_LDRS_PC_G2)
signed_value = value - pc + signed_addend;
else
signed_value = value - sb + signed_addend;
the residual at that stage. */
calculate_group_reloc_mask (abs (signed_value), group - 1, &residual);
if (residual >= 0x100)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): Overflow whilst splitting 0x%lx for group relocation %s"),
input_bfd, input_section,
(long) rel->r_offset, abs (signed_value), howto->name);
return bfd_reloc_overflow;
}
insn &= 0xff7ff0f0;
if (signed_value >= 0)
insn |= 1 << 23;
insn |= ((residual & 0xf0) << 4) | (residual & 0xf);
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
case R_ARM_LDC_PC_G0:
case R_ARM_LDC_PC_G1:
case R_ARM_LDC_PC_G2:
case R_ARM_LDC_SB_G0:
case R_ARM_LDC_SB_G1:
case R_ARM_LDC_SB_G2:
{
bfd_vma insn = bfd_get_32 (input_bfd, hit_data);
bfd_vma pc = input_section->output_section->vma
+ input_section->output_offset + rel->r_offset;
bfd_vma sb = 0;
bfd_vma residual;
bfd_signed_vma signed_value;
int group = 0;
switch (r_type)
{
case R_ARM_LDC_PC_G0:
case R_ARM_LDC_SB_G0:
group = 0;
break;
case R_ARM_LDC_PC_G1:
case R_ARM_LDC_SB_G1:
group = 1;
break;
case R_ARM_LDC_PC_G2:
case R_ARM_LDC_SB_G2:
group = 2;
break;
default:
abort ();
}
have already been fetched for us. */
if (globals->use_rel)
{
int negative = (insn & (1 << 23)) ? 1 : -1;
signed_addend = negative * ((insn & 0xff) << 2);
}
if (r_type == R_ARM_LDC_PC_G0
|| r_type == R_ARM_LDC_PC_G1
|| r_type == R_ARM_LDC_PC_G2)
signed_value = value - pc + signed_addend;
else
signed_value = value - sb + signed_addend;
the residual at that stage. */
calculate_group_reloc_mask (abs (signed_value), group - 1, &residual);
divisible by four and, after having been divided by four, must
fit in eight bits.) */
if ((residual & 0x3) != 0 || residual >= 0x400)
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): Overflow whilst splitting 0x%lx for group relocation %s"),
input_bfd, input_section,
(long) rel->r_offset, abs (signed_value), howto->name);
return bfd_reloc_overflow;
}
insn &= 0xff7fff00;
if (signed_value >= 0)
insn |= 1 << 23;
insn |= residual >> 2;
bfd_put_32 (input_bfd, insn, hit_data);
}
return bfd_reloc_ok;
default:
return bfd_reloc_notsupported;
}
}
static void
arm_add_to_rel (bfd * abfd,
bfd_byte * address,
reloc_howto_type * howto,
bfd_signed_vma increment)
{
bfd_signed_vma addend;
if (howto->type == R_ARM_THM_CALL
|| howto->type == R_ARM_THM_JUMP24)
{
int upper_insn, lower_insn;
int upper, lower;
upper_insn = bfd_get_16 (abfd, address);
lower_insn = bfd_get_16 (abfd, address + 2);
upper = upper_insn & 0x7ff;
lower = lower_insn & 0x7ff;
addend = (upper << 12) | (lower << 1);
addend += increment;
addend >>= 1;
upper_insn = (upper_insn & 0xf800) | ((addend >> 11) & 0x7ff);
lower_insn = (lower_insn & 0xf800) | (addend & 0x7ff);
bfd_put_16 (abfd, (bfd_vma) upper_insn, address);
bfd_put_16 (abfd, (bfd_vma) lower_insn, address + 2);
}
else
{
bfd_vma contents;
contents = bfd_get_32 (abfd, address);
addend = contents & howto->src_mask;
if (addend & ((howto->src_mask + 1) >> 1))
{
bfd_signed_vma mask;
mask = -1;
mask &= ~ howto->src_mask;
addend |= mask;
}
switch (howto->type)
{
default:
addend += increment;
break;
case R_ARM_PC24:
case R_ARM_PLT32:
case R_ARM_CALL:
case R_ARM_JUMP24:
addend <<= howto->size;
addend += increment;
addend >>= howto->rightshift;
break;
}
contents = (contents & ~ howto->dst_mask) | (addend & howto->dst_mask);
bfd_put_32 (abfd, contents, address);
}
}
#define IS_ARM_TLS_RELOC(R_TYPE) \
((R_TYPE) == R_ARM_TLS_GD32 \
|| (R_TYPE) == R_ARM_TLS_LDO32 \
|| (R_TYPE) == R_ARM_TLS_LDM32 \
|| (R_TYPE) == R_ARM_TLS_DTPOFF32 \
|| (R_TYPE) == R_ARM_TLS_DTPMOD32 \
|| (R_TYPE) == R_ARM_TLS_TPOFF32 \
|| (R_TYPE) == R_ARM_TLS_LE32 \
|| (R_TYPE) == R_ARM_TLS_IE32)
static bfd_boolean
elf32_arm_relocate_section (bfd * output_bfd,
struct bfd_link_info * info,
bfd * input_bfd,
asection * input_section,
bfd_byte * contents,
Elf_Internal_Rela * relocs,
Elf_Internal_Sym * local_syms,
asection ** local_sections)
{
Elf_Internal_Shdr *symtab_hdr;
struct elf_link_hash_entry **sym_hashes;
Elf_Internal_Rela *rel;
Elf_Internal_Rela *relend;
const char *name;
struct elf32_arm_link_hash_table * globals;
globals = elf32_arm_hash_table (info);
symtab_hdr = & elf_symtab_hdr (input_bfd);
sym_hashes = elf_sym_hashes (input_bfd);
rel = relocs;
relend = relocs + input_section->reloc_count;
for (; rel < relend; rel++)
{
int r_type;
reloc_howto_type * howto;
unsigned long r_symndx;
Elf_Internal_Sym * sym;
asection * sec;
struct elf_link_hash_entry * h;
bfd_vma relocation;
bfd_reloc_status_type r;
arelent bfd_reloc;
char sym_type;
bfd_boolean unresolved_reloc = FALSE;
char *error_message = NULL;
r_symndx = ELF32_R_SYM (rel->r_info);
r_type = ELF32_R_TYPE (rel->r_info);
r_type = arm_real_reloc_type (globals, r_type);
if ( r_type == R_ARM_GNU_VTENTRY
|| r_type == R_ARM_GNU_VTINHERIT)
continue;
bfd_reloc.howto = elf32_arm_howto_from_type (r_type);
howto = bfd_reloc.howto;
h = NULL;
sym = NULL;
sec = NULL;
if (r_symndx < symtab_hdr->sh_info)
{
sym = local_syms + r_symndx;
sym_type = ELF32_ST_TYPE (sym->st_info);
sec = local_sections[r_symndx];
if (globals->use_rel)
{
relocation = (sec->output_section->vma
+ sec->output_offset
+ sym->st_value);
if (!info->relocatable
&& (sec->flags & SEC_MERGE)
&& ELF_ST_TYPE (sym->st_info) == STT_SECTION)
{
asection *msec;
bfd_vma addend, value;
switch (r_type)
{
case R_ARM_MOVW_ABS_NC:
case R_ARM_MOVT_ABS:
value = bfd_get_32 (input_bfd, contents + rel->r_offset);
addend = ((value & 0xf0000) >> 4) | (value & 0xfff);
addend = (addend ^ 0x8000) - 0x8000;
break;
case R_ARM_THM_MOVW_ABS_NC:
case R_ARM_THM_MOVT_ABS:
value = bfd_get_16 (input_bfd, contents + rel->r_offset)
<< 16;
value |= bfd_get_16 (input_bfd,
contents + rel->r_offset + 2);
addend = ((value & 0xf7000) >> 4) | (value & 0xff)
| ((value & 0x04000000) >> 15);
addend = (addend ^ 0x8000) - 0x8000;
break;
default:
if (howto->rightshift
|| (howto->src_mask & (howto->src_mask + 1)))
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): %s relocation against SEC_MERGE section"),
input_bfd, input_section,
(long) rel->r_offset, howto->name);
return FALSE;
}
value = bfd_get_32 (input_bfd, contents + rel->r_offset);
addend = value & howto->src_mask;
if (addend & ((howto->src_mask + 1) >> 1))
{
bfd_signed_vma mask;
mask = -1;
mask &= ~ howto->src_mask;
addend |= mask;
}
break;
}
msec = sec;
addend =
_bfd_elf_rel_local_sym (output_bfd, sym, &msec, addend)
- relocation;
addend += msec->output_section->vma + msec->output_offset;
switch statement. */
switch (r_type)
{
case R_ARM_MOVW_ABS_NC:
case R_ARM_MOVT_ABS:
value = (value & 0xfff0f000) | ((addend & 0xf000) << 4)
| (addend & 0xfff);
bfd_put_32 (input_bfd, value, contents + rel->r_offset);
break;
case R_ARM_THM_MOVW_ABS_NC:
case R_ARM_THM_MOVT_ABS:
value = (value & 0xfbf08f00) | ((addend & 0xf700) << 4)
| (addend & 0xff) | ((addend & 0x0800) << 15);
bfd_put_16 (input_bfd, value >> 16,
contents + rel->r_offset);
bfd_put_16 (input_bfd, value,
contents + rel->r_offset + 2);
break;
default:
value = (value & ~ howto->dst_mask)
| (addend & howto->dst_mask);
bfd_put_32 (input_bfd, value, contents + rel->r_offset);
break;
}
}
}
else
relocation = _bfd_elf_rela_local_sym (output_bfd, sym, &sec, rel);
}
else
{
bfd_boolean warned;
RELOC_FOR_GLOBAL_SYMBOL (info, input_bfd, input_section, rel,
r_symndx, symtab_hdr, sym_hashes,
h, sec, relocation,
unresolved_reloc, warned);
sym_type = h->type;
}
if (sec != NULL && elf_discarded_section (sec))
{
or sections discarded by a linker script, we just want the
section contents zeroed. Avoid any special processing. */
_bfd_clear_contents (howto, input_bfd, contents + rel->r_offset);
rel->r_info = 0;
rel->r_addend = 0;
continue;
}
if (info->relocatable)
{
anything, unless the reloc is against a section symbol,
in which case we have to adjust according to where the
section symbol winds up in the output section. */
if (sym != NULL && ELF_ST_TYPE (sym->st_info) == STT_SECTION)
{
if (globals->use_rel)
arm_add_to_rel (input_bfd, contents + rel->r_offset,
howto, (bfd_signed_vma) sec->output_offset);
else
rel->r_addend += sec->output_offset;
}
continue;
}
if (h != NULL)
name = h->root.root.string;
else
{
name = (bfd_elf_string_from_elf_section
(input_bfd, symtab_hdr->sh_link, sym->st_name));
if (name == NULL || *name == '\0')
name = bfd_section_name (input_bfd, sec);
}
if (r_symndx != 0
&& r_type != R_ARM_NONE
&& (h == NULL
|| h->root.type == bfd_link_hash_defined
|| h->root.type == bfd_link_hash_defweak)
&& IS_ARM_TLS_RELOC (r_type) != (sym_type == STT_TLS))
{
(*_bfd_error_handler)
((sym_type == STT_TLS
? _("%B(%A+0x%lx): %s used with TLS symbol %s")
: _("%B(%A+0x%lx): %s used with non-TLS symbol %s")),
input_bfd,
input_section,
(long) rel->r_offset,
howto->name,
name);
}
r = elf32_arm_final_link_relocate (howto, input_bfd, output_bfd,
input_section, contents, rel,
relocation, info, sec, name,
(h ? ELF_ST_TYPE (h->type) :
ELF_ST_TYPE (sym->st_info)), h,
&unresolved_reloc, &error_message);
because such sections are not SEC_ALLOC and thus ld.so will
not process them. */
if (unresolved_reloc
&& !((input_section->flags & SEC_DEBUGGING) != 0
&& h->def_dynamic))
{
(*_bfd_error_handler)
(_("%B(%A+0x%lx): unresolvable %s relocation against symbol `%s'"),
input_bfd,
input_section,
(long) rel->r_offset,
howto->name,
h->root.root.string);
return FALSE;
}
if (r != bfd_reloc_ok)
{
switch (r)
{
case bfd_reloc_overflow:
we have already printed one error message and there
is no point complaining again. */
if ((! h ||
h->root.type != bfd_link_hash_undefined)
&& (!((*info->callbacks->reloc_overflow)
(info, (h ? &h->root : NULL), name, howto->name,
(bfd_vma) 0, input_bfd, input_section,
rel->r_offset))))
return FALSE;
break;
case bfd_reloc_undefined:
if (!((*info->callbacks->undefined_symbol)
(info, name, input_bfd, input_section,
rel->r_offset, TRUE)))
return FALSE;
break;
case bfd_reloc_outofrange:
error_message = _("out of range");
goto common_error;
case bfd_reloc_notsupported:
error_message = _("unsupported relocation");
goto common_error;
case bfd_reloc_dangerous:
goto common_error;
default:
error_message = _("unknown error");
common_error:
BFD_ASSERT (error_message != NULL);
if (!((*info->callbacks->reloc_dangerous)
(info, error_message, input_bfd, input_section,
rel->r_offset)))
return FALSE;
break;
}
}
}
return TRUE;
}
static bfd_boolean
elf32_arm_object_p (bfd *abfd)
{
unsigned int mach;
mach = bfd_arm_get_mach_from_notes (abfd, ARM_NOTE_SECTION);
if (mach != bfd_mach_arm_unknown)
bfd_default_set_arch_mach (abfd, bfd_arch_arm, mach);
else if (elf_elfheader (abfd)->e_flags & EF_ARM_MAVERICK_FLOAT)
bfd_default_set_arch_mach (abfd, bfd_arch_arm, bfd_mach_arm_ep9312);
else
bfd_default_set_arch_mach (abfd, bfd_arch_arm, mach);
return TRUE;
}
static bfd_boolean
elf32_arm_set_private_flags (bfd *abfd, flagword flags)
{
if (elf_flags_init (abfd)
&& elf_elfheader (abfd)->e_flags != flags)
{
if (EF_ARM_EABI_VERSION (flags) == EF_ARM_EABI_UNKNOWN)
{
if (flags & EF_ARM_INTERWORK)
(*_bfd_error_handler)
(_("Warning: Not setting interworking flag of %B since it has already been specified as non-interworking"),
abfd);
else
_bfd_error_handler
(_("Warning: Clearing the interworking flag of %B due to outside request"),
abfd);
}
}
else
{
elf_elfheader (abfd)->e_flags = flags;
elf_flags_init (abfd) = TRUE;
}
return TRUE;
}
static bfd_boolean
elf32_arm_copy_private_bfd_data (bfd *ibfd, bfd *obfd)
{
flagword in_flags;
flagword out_flags;
if (! is_arm_elf (ibfd) || ! is_arm_elf (obfd))
return TRUE;
in_flags = elf_elfheader (ibfd)->e_flags;
out_flags = elf_elfheader (obfd)->e_flags;
if (elf_flags_init (obfd)
&& EF_ARM_EABI_VERSION (out_flags) == EF_ARM_EABI_UNKNOWN
&& in_flags != out_flags)
{
if ((in_flags & EF_ARM_APCS_26) != (out_flags & EF_ARM_APCS_26))
return FALSE;
if ((in_flags & EF_ARM_APCS_FLOAT) != (out_flags & EF_ARM_APCS_FLOAT))
return FALSE;
then turn off the interworking bit. */
if ((in_flags & EF_ARM_INTERWORK) != (out_flags & EF_ARM_INTERWORK))
{
if (out_flags & EF_ARM_INTERWORK)
_bfd_error_handler
(_("Warning: Clearing the interworking flag of %B because non-interworking code in %B has been linked with it"),
obfd, ibfd);
in_flags &= ~EF_ARM_INTERWORK;
}
if ((in_flags & EF_ARM_PIC) != (out_flags & EF_ARM_PIC))
in_flags &= ~EF_ARM_PIC;
}
elf_elfheader (obfd)->e_flags = in_flags;
elf_flags_init (obfd) = TRUE;
elf_elfheader (obfd)->e_ident[EI_OSABI] =
elf_elfheader (ibfd)->e_ident[EI_OSABI];
_bfd_elf_copy_obj_attributes (ibfd, obfd);
return TRUE;
}
enum
{
AEABI_R9_V6,
AEABI_R9_SB,
AEABI_R9_TLS,
AEABI_R9_unused
};
enum
{
AEABI_PCS_RW_data_absolute,
AEABI_PCS_RW_data_PCrel,
AEABI_PCS_RW_data_SBrel,
AEABI_PCS_RW_data_unused
};
enum
{
AEABI_enum_unused,
AEABI_enum_short,
AEABI_enum_wide,
AEABI_enum_forced_wide
};
string or both. */
static int
elf32_arm_obj_attrs_arg_type (int tag)
{
if (tag == Tag_compatibility)
return 3;
else if (tag == 4 || tag == 5)
return 2;
else if (tag < 32)
return 1;
else
return (tag & 1) != 0 ? 2 : 1;
}
are conflicting attributes. */
static bfd_boolean
elf32_arm_merge_eabi_attributes (bfd *ibfd, bfd *obfd)
{
obj_attribute *in_attr;
obj_attribute *out_attr;
obj_attribute_list *in_list;
2 = weak requirement. */
static const int order_312[3] = {3, 1, 2};
static const int order_01243[5] = {0, 1, 2, 4, 3};
int i;
if (!elf_known_obj_attributes_proc (obfd)[0].i)
{
_bfd_elf_copy_obj_attributes (ibfd, obfd);
initialized. */
elf_known_obj_attributes_proc (obfd)[0].i = 1;
return TRUE;
}
in_attr = elf_known_obj_attributes_proc (ibfd);
out_attr = elf_known_obj_attributes_proc (obfd);
if (in_attr[Tag_ABI_VFP_args].i != out_attr[Tag_ABI_VFP_args].i)
{
if (out_attr[Tag_ABI_FP_number_model].i == 0)
out_attr[Tag_ABI_VFP_args].i = in_attr[Tag_ABI_VFP_args].i;
else if (in_attr[Tag_ABI_FP_number_model].i != 0)
{
_bfd_error_handler
(_("ERROR: %B uses VFP register arguments, %B does not"),
ibfd, obfd);
return FALSE;
}
}
for (i = 4; i < NUM_KNOWN_OBJ_ATTRIBUTES; i++)
{
switch (i)
{
case Tag_CPU_raw_name:
case Tag_CPU_name:
won't necessarily have both the above tags, so make sure input
name is non-NULL. */
if (in_attr[Tag_CPU_arch].i > out_attr[Tag_CPU_arch].i
&& in_attr[i].s)
out_attr[i].s = _bfd_elf_attr_strdup (obfd, in_attr[i].s);
break;
case Tag_ABI_optimization_goals:
case Tag_ABI_FP_optimization_goals:
break;
case Tag_CPU_arch:
case Tag_ARM_ISA_use:
case Tag_THUMB_ISA_use:
case Tag_WMMX_arch:
case Tag_NEON_arch:
case Tag_ABI_FP_rounding:
case Tag_ABI_FP_denormal:
case Tag_ABI_FP_exceptions:
case Tag_ABI_FP_user_exceptions:
case Tag_ABI_FP_number_model:
case Tag_ABI_align8_preserved:
case Tag_ABI_HardFP_use:
if (in_attr[i].i > out_attr[i].i)
out_attr[i].i = in_attr[i].i;
break;
case Tag_CPU_arch_profile:
if (out_attr[i].i && in_attr[i].i && in_attr[i].i != out_attr[i].i)
{
_bfd_error_handler
(_("ERROR: %B: Conflicting architecture profiles %c/%c"),
ibfd, in_attr[i].i, out_attr[i].i);
return FALSE;
}
if (in_attr[i].i)
out_attr[i].i = in_attr[i].i;
break;
case Tag_VFP_arch:
if (in_attr[i].i > 4 || out_attr[i].i > 4
|| order_01243[in_attr[i].i] > order_01243[out_attr[i].i])
out_attr[i].i = in_attr[i].i;
break;
case Tag_PCS_config:
if (out_attr[i].i == 0)
out_attr[i].i = in_attr[i].i;
else if (in_attr[i].i != 0 && out_attr[i].i != 0)
{
a warning. */
_bfd_error_handler
(_("Warning: %B: Conflicting platform configuration"), ibfd);
}
break;
case Tag_ABI_PCS_R9_use:
if (in_attr[i].i != out_attr[i].i
&& out_attr[i].i != AEABI_R9_unused
&& in_attr[i].i != AEABI_R9_unused)
{
_bfd_error_handler
(_("ERROR: %B: Conflicting use of R9"), ibfd);
return FALSE;
}
if (out_attr[i].i == AEABI_R9_unused)
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_PCS_RW_data:
if (in_attr[i].i == AEABI_PCS_RW_data_SBrel
&& out_attr[Tag_ABI_PCS_R9_use].i != AEABI_R9_SB
&& out_attr[Tag_ABI_PCS_R9_use].i != AEABI_R9_unused)
{
_bfd_error_handler
(_("ERROR: %B: SB relative addressing conflicts with use of R9"),
ibfd);
return FALSE;
}
if (in_attr[i].i < out_attr[i].i)
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_PCS_RO_data:
if (in_attr[i].i < out_attr[i].i)
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_PCS_GOT_use:
if (in_attr[i].i > 2 || out_attr[i].i > 2
|| order_312[in_attr[i].i] < order_312[out_attr[i].i])
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_PCS_wchar_t:
if (out_attr[i].i && in_attr[i].i && out_attr[i].i != in_attr[i].i
&& !elf_arm_tdata (obfd)->no_wchar_size_warning)
{
_bfd_error_handler
(_("warning: %B uses %u-byte wchar_t yet the output is to use %u-byte wchar_t; use of wchar_t values across objects may fail"),
ibfd, in_attr[i].i, out_attr[i].i);
}
else if (in_attr[i].i && !out_attr[i].i)
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_align8_needed:
if (in_attr[i].i > 2 || out_attr[i].i > 2
|| order_312[in_attr[i].i] < order_312[out_attr[i].i])
out_attr[i].i = in_attr[i].i;
break;
case Tag_ABI_enum_size:
if (in_attr[i].i != AEABI_enum_unused)
{
if (out_attr[i].i == AEABI_enum_unused
|| out_attr[i].i == AEABI_enum_forced_wide)
{
Use whatever requirements the new object has. */
out_attr[i].i = in_attr[i].i;
}
else if (in_attr[i].i != AEABI_enum_forced_wide
&& out_attr[i].i != in_attr[i].i
&& !elf_arm_tdata (obfd)->no_enum_size_warning)
{
const char *aeabi_enum_names[] =
{ "", "variable-size", "32-bit", "" };
_bfd_error_handler
(_("warning: %B uses %s enums yet the output is to use %s enums; use of enum values across objects may fail"),
ibfd, aeabi_enum_names[in_attr[i].i],
aeabi_enum_names[out_attr[i].i]);
}
}
break;
case Tag_ABI_VFP_args:
break;
case Tag_ABI_WMMX_args:
if (in_attr[i].i != out_attr[i].i)
{
_bfd_error_handler
(_("ERROR: %B uses iWMMXt register arguments, %B does not"),
ibfd, obfd);
return FALSE;
}
break;
default:
abort ();
}
if (in_attr[i].type && !out_attr[i].type)
switch (in_attr[i].type)
{
case 1:
if (out_attr[i].i)
out_attr[i].type = 1;
break;
case 2:
if (out_attr[i].s)
out_attr[i].type = 2;
break;
default:
abort ();
}
}
_bfd_elf_merge_object_attributes (ibfd, obfd);
in_list = elf_other_obj_attributes_proc (ibfd);
while (in_list && in_list->tag == Tag_compatibility)
in_list = in_list->next;
for (; in_list; in_list = in_list->next)
{
if ((in_list->tag & 128) < 64)
{
_bfd_error_handler
(_("Warning: %B: Unknown EABI object attribute %d"),
ibfd, in_list->tag);
break;
}
}
return TRUE;
}
static bfd_boolean
elf32_arm_versions_compatible (unsigned iver, unsigned over)
{
so allow mixing them. */
if ((iver == EF_ARM_EABI_VER4 && over == EF_ARM_EABI_VER5)
|| (iver == EF_ARM_EABI_VER5 && over == EF_ARM_EABI_VER4))
return TRUE;
return (iver == over);
}
object file when linking. */
static bfd_boolean
elf32_arm_merge_private_bfd_data (bfd * ibfd, bfd * obfd)
{
flagword out_flags;
flagword in_flags;
bfd_boolean flags_compatible = TRUE;
asection *sec;
if (! _bfd_generic_verify_endian_match (ibfd, obfd))
return FALSE;
if (! is_arm_elf (ibfd) || ! is_arm_elf (obfd))
return TRUE;
if (!elf32_arm_merge_eabi_attributes (ibfd, obfd))
return FALSE;
the assembler but I don't think an elf_flags_init field is
written into the object. */
in_flags = elf_elfheader (ibfd)->e_flags;
out_flags = elf_elfheader (obfd)->e_flags;
in practice it isn't even close to working and there is no real
reason to want it. */
if (EF_ARM_EABI_VERSION (in_flags) >= EF_ARM_EABI_VER4
&& !(ibfd->flags & DYNAMIC)
&& (in_flags & EF_ARM_BE8))
{
_bfd_error_handler (_("ERROR: %B is already in final BE8 format"),
ibfd);
return FALSE;
}
if (!elf_flags_init (obfd))
{
flags then do not bother setting the flags for the output
architecture, instead allow future merges to do this. If no
future merges ever set these flags then they will retain their
uninitialised values, which surprise surprise, correspond
to the default values. */
if (bfd_get_arch_info (ibfd)->the_default
&& elf_elfheader (ibfd)->e_flags == 0)
return TRUE;
elf_flags_init (obfd) = TRUE;
elf_elfheader (obfd)->e_flags = in_flags;
if (bfd_get_arch (obfd) == bfd_get_arch (ibfd)
&& bfd_get_arch_info (obfd)->the_default)
return bfd_set_arch_mach (obfd, bfd_get_arch (ibfd), bfd_get_mach (ibfd));
return TRUE;
}
does not match the output ARM architecture. */
if (! bfd_arm_merge_machines (ibfd, obfd))
return FALSE;
if (in_flags == out_flags)
return TRUE;
not, its flags may not have been initialised either, but it
cannot actually cause any incompatiblity. Do not short-circuit
dynamic objects; their section list may be emptied by
elf_link_add_object_symbols.
Also check to see if there are no code sections in the input.
In this case there is no need to check for code specific flags.
XXX - do we need to worry about floating-point format compatability
in data sections ? */
if (!(ibfd->flags & DYNAMIC))
{
bfd_boolean null_input_bfd = TRUE;
bfd_boolean only_data_sections = TRUE;
for (sec = ibfd->sections; sec != NULL; sec = sec->next)
{
if (strcmp (sec->name, ".glue_7")
&& strcmp (sec->name, ".glue_7t"))
{
if ((bfd_get_section_flags (ibfd, sec)
& (SEC_LOAD | SEC_CODE | SEC_HAS_CONTENTS))
== (SEC_LOAD | SEC_CODE | SEC_HAS_CONTENTS))
only_data_sections = FALSE;
null_input_bfd = FALSE;
break;
}
}
if (null_input_bfd || only_data_sections)
return TRUE;
}
if (!elf32_arm_versions_compatible (EF_ARM_EABI_VERSION (in_flags),
EF_ARM_EABI_VERSION (out_flags)))
{
_bfd_error_handler
(_("ERROR: Source object %B has EABI version %d, but target %B has EABI version %d"),
ibfd, obfd,
(in_flags & EF_ARM_EABIMASK) >> 24,
(out_flags & EF_ARM_EABIMASK) >> 24);
return FALSE;
}
if (get_elf_backend_data (obfd) != &elf32_arm_vxworks_bed
&& get_elf_backend_data (ibfd) != &elf32_arm_vxworks_bed
&& EF_ARM_EABI_VERSION (in_flags) == EF_ARM_EABI_UNKNOWN)
{
if ((in_flags & EF_ARM_APCS_26) != (out_flags & EF_ARM_APCS_26))
{
_bfd_error_handler
(_("ERROR: %B is compiled for APCS-%d, whereas target %B uses APCS-%d"),
ibfd, obfd,
in_flags & EF_ARM_APCS_26 ? 26 : 32,
out_flags & EF_ARM_APCS_26 ? 26 : 32);
flags_compatible = FALSE;
}
if ((in_flags & EF_ARM_APCS_FLOAT) != (out_flags & EF_ARM_APCS_FLOAT))
{
if (in_flags & EF_ARM_APCS_FLOAT)
_bfd_error_handler
(_("ERROR: %B passes floats in float registers, whereas %B passes them in integer registers"),
ibfd, obfd);
else
_bfd_error_handler
(_("ERROR: %B passes floats in integer registers, whereas %B passes them in float registers"),
ibfd, obfd);
flags_compatible = FALSE;
}
if ((in_flags & EF_ARM_VFP_FLOAT) != (out_flags & EF_ARM_VFP_FLOAT))
{
if (in_flags & EF_ARM_VFP_FLOAT)
_bfd_error_handler
(_("ERROR: %B uses VFP instructions, whereas %B does not"),
ibfd, obfd);
else
_bfd_error_handler
(_("ERROR: %B uses FPA instructions, whereas %B does not"),
ibfd, obfd);
flags_compatible = FALSE;
}
if ((in_flags & EF_ARM_MAVERICK_FLOAT) != (out_flags & EF_ARM_MAVERICK_FLOAT))
{
if (in_flags & EF_ARM_MAVERICK_FLOAT)
_bfd_error_handler
(_("ERROR: %B uses Maverick instructions, whereas %B does not"),
ibfd, obfd);
else
_bfd_error_handler
(_("ERROR: %B does not use Maverick instructions, whereas %B does"),
ibfd, obfd);
flags_compatible = FALSE;
}
#ifdef EF_ARM_SOFT_FLOAT
if ((in_flags & EF_ARM_SOFT_FLOAT) != (out_flags & EF_ARM_SOFT_FLOAT))
{
layout, and uses either soft float or integer regs for
passing floating point arguments and results. We already
know that the APCS_FLOAT flags match; similarly for VFP
flags. */
if ((in_flags & EF_ARM_APCS_FLOAT) != 0
|| (in_flags & EF_ARM_VFP_FLOAT) == 0)
{
if (in_flags & EF_ARM_SOFT_FLOAT)
_bfd_error_handler
(_("ERROR: %B uses software FP, whereas %B uses hardware FP"),
ibfd, obfd);
else
_bfd_error_handler
(_("ERROR: %B uses hardware FP, whereas %B uses software FP"),
ibfd, obfd);
flags_compatible = FALSE;
}
}
#endif
if ((in_flags & EF_ARM_INTERWORK) != (out_flags & EF_ARM_INTERWORK))
{
if (in_flags & EF_ARM_INTERWORK)
{
_bfd_error_handler
(_("Warning: %B supports interworking, whereas %B does not"),
ibfd, obfd);
}
else
{
_bfd_error_handler
(_("Warning: %B does not support interworking, whereas %B does"),
ibfd, obfd);
}
}
}
return flags_compatible;
}
static bfd_boolean
elf32_arm_print_private_bfd_data (bfd *abfd, void * ptr)
{
FILE * file = (FILE *) ptr;
unsigned long flags;
BFD_ASSERT (abfd != NULL && ptr != NULL);
_bfd_elf_print_private_bfd_data (abfd, ptr);
flags = elf_elfheader (abfd)->e_flags;
containing valid data. */
fprintf (file, _("private flags = %lx:"), elf_elfheader (abfd)->e_flags);
switch (EF_ARM_EABI_VERSION (flags))
{
case EF_ARM_EABI_UNKNOWN:
official ARM ELF extended ABI. Hence they are only decoded if
the EABI version is not set. */
if (flags & EF_ARM_INTERWORK)
fprintf (file, _(" [interworking enabled]"));
if (flags & EF_ARM_APCS_26)
fprintf (file, " [APCS-26]");
else
fprintf (file, " [APCS-32]");
if (flags & EF_ARM_VFP_FLOAT)
fprintf (file, _(" [VFP float format]"));
else if (flags & EF_ARM_MAVERICK_FLOAT)
fprintf (file, _(" [Maverick float format]"));
else
fprintf (file, _(" [FPA float format]"));
if (flags & EF_ARM_APCS_FLOAT)
fprintf (file, _(" [floats passed in float registers]"));
if (flags & EF_ARM_PIC)
fprintf (file, _(" [position independent]"));
if (flags & EF_ARM_NEW_ABI)
fprintf (file, _(" [new ABI]"));
if (flags & EF_ARM_OLD_ABI)
fprintf (file, _(" [old ABI]"));
if (flags & EF_ARM_SOFT_FLOAT)
fprintf (file, _(" [software FP]"));
flags &= ~(EF_ARM_INTERWORK | EF_ARM_APCS_26 | EF_ARM_APCS_FLOAT
| EF_ARM_PIC | EF_ARM_NEW_ABI | EF_ARM_OLD_ABI
| EF_ARM_SOFT_FLOAT | EF_ARM_VFP_FLOAT
| EF_ARM_MAVERICK_FLOAT);
break;
case EF_ARM_EABI_VER1:
fprintf (file, _(" [Version1 EABI]"));
if (flags & EF_ARM_SYMSARESORTED)
fprintf (file, _(" [sorted symbol table]"));
else
fprintf (file, _(" [unsorted symbol table]"));
flags &= ~ EF_ARM_SYMSARESORTED;
break;
case EF_ARM_EABI_VER2:
fprintf (file, _(" [Version2 EABI]"));
if (flags & EF_ARM_SYMSARESORTED)
fprintf (file, _(" [sorted symbol table]"));
else
fprintf (file, _(" [unsorted symbol table]"));
if (flags & EF_ARM_DYNSYMSUSESEGIDX)
fprintf (file, _(" [dynamic symbols use segment index]"));
if (flags & EF_ARM_MAPSYMSFIRST)
fprintf (file, _(" [mapping symbols precede others]"));
flags &= ~(EF_ARM_SYMSARESORTED | EF_ARM_DYNSYMSUSESEGIDX
| EF_ARM_MAPSYMSFIRST);
break;
case EF_ARM_EABI_VER3:
fprintf (file, _(" [Version3 EABI]"));
break;
case EF_ARM_EABI_VER4:
fprintf (file, _(" [Version4 EABI]"));
goto eabi;
case EF_ARM_EABI_VER5:
fprintf (file, _(" [Version5 EABI]"));
eabi:
if (flags & EF_ARM_BE8)
fprintf (file, _(" [BE8]"));
if (flags & EF_ARM_LE8)
fprintf (file, _(" [LE8]"));
flags &= ~(EF_ARM_LE8 | EF_ARM_BE8);
break;
default:
fprintf (file, _(" <EABI version unrecognised>"));
break;
}
flags &= ~ EF_ARM_EABIMASK;
if (flags & EF_ARM_RELEXEC)
fprintf (file, _(" [relocatable executable]"));
if (flags & EF_ARM_HASENTRY)
fprintf (file, _(" [has entry point]"));
flags &= ~ (EF_ARM_RELEXEC | EF_ARM_HASENTRY);
if (flags)
fprintf (file, _("<Unrecognised flag bits set>"));
fputc ('\n', file);
return TRUE;
}
static int
elf32_arm_get_symbol_type (Elf_Internal_Sym * elf_sym, int type)
{
switch (ELF_ST_TYPE (elf_sym->st_info))
{
case STT_ARM_TFUNC:
return ELF_ST_TYPE (elf_sym->st_info);
case STT_ARM_16BIT:
This allows us to distinguish between data used by Thumb instructions
and non-data (which is probably code) inside Thumb regions of an
executable. */
if (type != STT_OBJECT && type != STT_TLS)
return ELF_ST_TYPE (elf_sym->st_info);
break;
default:
break;
}
return type;
}
static asection *
elf32_arm_gc_mark_hook (asection *sec,
struct bfd_link_info *info,
Elf_Internal_Rela *rel,
struct elf_link_hash_entry *h,
Elf_Internal_Sym *sym)
{
if (h != NULL)
switch (ELF32_R_TYPE (rel->r_info))
{
case R_ARM_GNU_VTINHERIT:
case R_ARM_GNU_VTENTRY:
return NULL;
}
return _bfd_elf_gc_mark_hook (sec, info, rel, h, sym);
}
static bfd_boolean
elf32_arm_gc_sweep_hook (bfd * abfd,
struct bfd_link_info * info,
asection * sec,
const Elf_Internal_Rela * relocs)
{
Elf_Internal_Shdr *symtab_hdr;
struct elf_link_hash_entry **sym_hashes;
bfd_signed_vma *local_got_refcounts;
const Elf_Internal_Rela *rel, *relend;
struct elf32_arm_link_hash_table * globals;
if (info->relocatable)
return TRUE;
globals = elf32_arm_hash_table (info);
elf_section_data (sec)->local_dynrel = NULL;
symtab_hdr = & elf_symtab_hdr (abfd);
sym_hashes = elf_sym_hashes (abfd);
local_got_refcounts = elf_local_got_refcounts (abfd);
check_use_blx (globals);
relend = relocs + sec->reloc_count;
for (rel = relocs; rel < relend; rel++)
{
unsigned long r_symndx;
struct elf_link_hash_entry *h = NULL;
int r_type;
r_symndx = ELF32_R_SYM (rel->r_info);
if (r_symndx >= symtab_hdr->sh_info)
{
h = sym_hashes[r_symndx - symtab_hdr->sh_info];
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
}
r_type = ELF32_R_TYPE (rel->r_info);
r_type = arm_real_reloc_type (globals, r_type);
switch (r_type)
{
case R_ARM_GOT32:
case R_ARM_GOT_PREL:
case R_ARM_TLS_GD32:
case R_ARM_TLS_IE32:
if (h != NULL)
{
if (h->got.refcount > 0)
h->got.refcount -= 1;
}
else if (local_got_refcounts != NULL)
{
if (local_got_refcounts[r_symndx] > 0)
local_got_refcounts[r_symndx] -= 1;
}
break;
case R_ARM_TLS_LDM32:
elf32_arm_hash_table (info)->tls_ldm_got.refcount -= 1;
break;
case R_ARM_ABS32:
case R_ARM_ABS32_NOI:
case R_ARM_REL32:
case R_ARM_REL32_NOI:
case R_ARM_PC24:
case R_ARM_PLT32:
case R_ARM_CALL:
case R_ARM_JUMP24:
case R_ARM_PREL31:
case R_ARM_THM_CALL:
case R_ARM_THM_JUMP24:
case R_ARM_THM_JUMP19:
case R_ARM_MOVW_ABS_NC:
case R_ARM_MOVT_ABS:
case R_ARM_MOVW_PREL_NC:
case R_ARM_MOVT_PREL:
case R_ARM_THM_MOVW_ABS_NC:
case R_ARM_THM_MOVT_ABS:
case R_ARM_THM_MOVW_PREL_NC:
case R_ARM_THM_MOVT_PREL:
if (h != NULL)
{
struct elf32_arm_link_hash_entry *eh;
struct elf32_arm_relocs_copied **pp;
struct elf32_arm_relocs_copied *p;
eh = (struct elf32_arm_link_hash_entry *) h;
if (h->plt.refcount > 0)
{
h->plt.refcount -= 1;
if (r_type == R_ARM_THM_CALL)
eh->plt_maybe_thumb_refcount--;
if (r_type == R_ARM_THM_JUMP24
|| r_type == R_ARM_THM_JUMP19)
eh->plt_thumb_refcount--;
}
if (r_type == R_ARM_ABS32
|| r_type == R_ARM_REL32
|| r_type == R_ARM_ABS32_NOI
|| r_type == R_ARM_REL32_NOI)
{
for (pp = &eh->relocs_copied; (p = *pp) != NULL;
pp = &p->next)
if (p->section == sec)
{
p->count -= 1;
if (ELF32_R_TYPE (rel->r_info) == R_ARM_REL32
|| ELF32_R_TYPE (rel->r_info) == R_ARM_REL32_NOI)
p->pc_count -= 1;
if (p->count == 0)
*pp = p->next;
break;
}
}
}
break;
default:
break;
}
}
return TRUE;
}
static bfd_boolean
elf32_arm_check_relocs (bfd *abfd, struct bfd_link_info *info,
asection *sec, const Elf_Internal_Rela *relocs)
{
Elf_Internal_Shdr *symtab_hdr;
struct elf_link_hash_entry **sym_hashes;
const Elf_Internal_Rela *rel;
const Elf_Internal_Rela *rel_end;
bfd *dynobj;
asection *sreloc;
bfd_vma *local_got_offsets;
struct elf32_arm_link_hash_table *htab;
bfd_boolean needs_plt;
if (info->relocatable)
return TRUE;
BFD_ASSERT (is_arm_elf (abfd));
htab = elf32_arm_hash_table (info);
sreloc = NULL;
copy relocations. */
if (htab->root.is_relocatable_executable
&& ! htab->root.dynamic_sections_created)
{
if (! _bfd_elf_link_create_dynamic_sections (abfd, info))
return FALSE;
}
dynobj = elf_hash_table (info)->dynobj;
local_got_offsets = elf_local_got_offsets (abfd);
symtab_hdr = & elf_symtab_hdr (abfd);
sym_hashes = elf_sym_hashes (abfd);
rel_end = relocs + sec->reloc_count;
for (rel = relocs; rel < rel_end; rel++)
{
struct elf_link_hash_entry *h;
struct elf32_arm_link_hash_entry *eh;
unsigned long r_symndx;
int r_type;
r_symndx = ELF32_R_SYM (rel->r_info);
r_type = ELF32_R_TYPE (rel->r_info);
r_type = arm_real_reloc_type (htab, r_type);
if (r_symndx >= NUM_SHDR_ENTRIES (symtab_hdr))
{
(*_bfd_error_handler) (_("%B: bad symbol index: %d"), abfd,
r_symndx);
return FALSE;
}
if (r_symndx < symtab_hdr->sh_info)
h = NULL;
else
{
h = sym_hashes[r_symndx - symtab_hdr->sh_info];
while (h->root.type == bfd_link_hash_indirect
|| h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
}
eh = (struct elf32_arm_link_hash_entry *) h;
switch (r_type)
{
case R_ARM_GOT32:
case R_ARM_GOT_PREL:
case R_ARM_TLS_GD32:
case R_ARM_TLS_IE32:
{
int tls_type, old_tls_type;
switch (r_type)
{
case R_ARM_TLS_GD32: tls_type = GOT_TLS_GD; break;
case R_ARM_TLS_IE32: tls_type = GOT_TLS_IE; break;
default: tls_type = GOT_NORMAL; break;
}
if (h != NULL)
{
h->got.refcount++;
old_tls_type = elf32_arm_hash_entry (h)->tls_type;
}
else
{
bfd_signed_vma *local_got_refcounts;
local_got_refcounts = elf_local_got_refcounts (abfd);
if (local_got_refcounts == NULL)
{
bfd_size_type size;
size = symtab_hdr->sh_info;
size *= (sizeof (bfd_signed_vma) + sizeof (char));
local_got_refcounts = bfd_zalloc (abfd, size);
if (local_got_refcounts == NULL)
return FALSE;
elf_local_got_refcounts (abfd) = local_got_refcounts;
elf32_arm_local_got_tls_type (abfd)
= (char *) (local_got_refcounts + symtab_hdr->sh_info);
}
local_got_refcounts[r_symndx] += 1;
old_tls_type = elf32_arm_local_got_tls_type (abfd) [r_symndx];
}
TLS / non-TLS mismatch, based on the symbol type. We don't
support any linker relaxations. So just combine any TLS
types needed. */
if (old_tls_type != GOT_UNKNOWN && old_tls_type != GOT_NORMAL
&& tls_type != GOT_NORMAL)
tls_type |= old_tls_type;
if (old_tls_type != tls_type)
{
if (h != NULL)
elf32_arm_hash_entry (h)->tls_type = tls_type;
else
elf32_arm_local_got_tls_type (abfd) [r_symndx] = tls_type;
}
}
case R_ARM_TLS_LDM32:
if (r_type == R_ARM_TLS_LDM32)
htab->tls_ldm_got.refcount++;
case R_ARM_GOTOFF32:
case R_ARM_GOTPC:
if (htab->sgot == NULL)
{
if (htab->root.dynobj == NULL)
htab->root.dynobj = abfd;
if (!create_got_section (htab->root.dynobj, info))
return FALSE;
}
break;
case R_ARM_ABS12:
ldr __GOTT_INDEX__ offsets. */
if (!htab->vxworks_p)
break;
case R_ARM_PC24:
case R_ARM_PLT32:
case R_ARM_CALL:
case R_ARM_JUMP24:
case R_ARM_PREL31:
case R_ARM_THM_CALL:
case R_ARM_THM_JUMP24:
case R_ARM_THM_JUMP19:
needs_plt = 1;
goto normal_reloc;
case R_ARM_ABS32:
case R_ARM_ABS32_NOI:
case R_ARM_REL32:
case R_ARM_REL32_NOI:
case R_ARM_MOVW_ABS_NC:
case R_ARM_MOVT_ABS:
case R_ARM_MOVW_PREL_NC:
case R_ARM_MOVT_PREL:
case R_ARM_THM_MOVW_ABS_NC:
case R_ARM_THM_MOVT_ABS:
case R_ARM_THM_MOVW_PREL_NC:
case R_ARM_THM_MOVT_PREL:
needs_plt = 0;
normal_reloc:
if (h != NULL)
{
need a copy reloc. We can't check reliably at this
stage whether the section is read-only, as input
sections have not yet been mapped to output sections.
Tentatively set the flag for now, and correct in
adjust_dynamic_symbol. */
if (!info->shared)
h->non_got_ref = 1;
refers to is in a different object. We can't tell for
sure yet, because something later might force the
symbol local. */
if (needs_plt)
h->needs_plt = 1;
it, even if it's an ABS32 relocation. */
h->plt.refcount += 1;
record possible blx references separately from
relocs that definitely need a thumb stub. */
if (r_type == R_ARM_THM_CALL)
eh->plt_maybe_thumb_refcount += 1;
if (r_type == R_ARM_THM_JUMP24
|| r_type == R_ARM_THM_JUMP19)
eh->plt_thumb_refcount += 1;
}
and this is a reloc against a global symbol, or a non PC
relative reloc against a local symbol, then we need to copy
the reloc into the shared library. However, if we are linking
with -Bsymbolic, we do not need to copy a reloc against a
global symbol which is defined in an object we are
including in the link (i.e., DEF_REGULAR is set). At
this point we have not seen all the input files, so it is
possible that DEF_REGULAR is not set now but will be set
later (it is never cleared). We account for that
possibility below by storing information in the
relocs_copied field of the hash table entry. */
if ((info->shared || htab->root.is_relocatable_executable)
&& (sec->flags & SEC_ALLOC) != 0
&& ((r_type == R_ARM_ABS32 || r_type == R_ARM_ABS32_NOI)
|| (h != NULL && ! h->needs_plt
&& (! info->symbolic || ! h->def_regular))))
{
struct elf32_arm_relocs_copied *p, **head;
reloc types into the output file. We create a reloc
section in dynobj and make room for this reloc. */
if (sreloc == NULL)
{
const char * name;
name = (bfd_elf_string_from_elf_section
(abfd,
elf_elfheader (abfd)->e_shstrndx,
elf_section_data (sec)->rel_hdr.sh_name));
if (name == NULL)
return FALSE;
BFD_ASSERT (reloc_section_p (htab, name, sec));
sreloc = bfd_get_section_by_name (dynobj, name);
if (sreloc == NULL)
{
flagword flags;
flags = (SEC_HAS_CONTENTS | SEC_READONLY
| SEC_IN_MEMORY | SEC_LINKER_CREATED);
if ((sec->flags & SEC_ALLOC) != 0
relocations mapped. */
&& !htab->symbian_p)
flags |= SEC_ALLOC | SEC_LOAD;
sreloc = bfd_make_section_with_flags (dynobj,
name,
flags);
if (sreloc == NULL
|| ! bfd_set_section_alignment (dynobj, sreloc, 2))
return FALSE;
}
elf_section_data (sec)->sreloc = sreloc;
}
relocations we need for this symbol. */
if (h != NULL)
{
head = &((struct elf32_arm_link_hash_entry *) h)->relocs_copied;
}
else
{
We really need local syms available to do this
easily. Oh well. */
asection *s;
void *vpp;
s = bfd_section_from_r_symndx (abfd, &htab->sym_sec,
sec, r_symndx);
if (s == NULL)
return FALSE;
vpp = &elf_section_data (s)->local_dynrel;
head = (struct elf32_arm_relocs_copied **) vpp;
}
p = *head;
if (p == NULL || p->section != sec)
{
bfd_size_type amt = sizeof *p;
p = bfd_alloc (htab->root.dynobj, amt);
if (p == NULL)
return FALSE;
p->next = *head;
*head = p;
p->section = sec;
p->count = 0;
p->pc_count = 0;
}
if (r_type == R_ARM_REL32 || r_type == R_ARM_REL32_NOI)
p->pc_count += 1;
p->count += 1;
}
break;
Reconstruct it for later use during GC. */
case R_ARM_GNU_VTINHERIT:
if (!bfd_elf_gc_record_vtinherit (abfd, sec, h, rel->r_offset))
return FALSE;
break;
used. Record for later use during GC. */
case R_ARM_GNU_VTENTRY:
BFD_ASSERT (h != NULL);
if (h != NULL
&& !bfd_elf_gc_record_vtentry (abfd, sec, h, rel->r_offset))
return FALSE;
break;
}
}
return TRUE;
}
required if the corresponding code section is marked. */
static bfd_boolean
elf32_arm_gc_mark_extra_sections (struct bfd_link_info *info,
elf_gc_mark_hook_fn gc_mark_hook)
{
bfd *sub;
Elf_Internal_Shdr **elf_shdrp;
bfd_boolean again;
requiring multiple passes. */
again = TRUE;
while (again)
{
again = FALSE;
for (sub = info->input_bfds; sub != NULL; sub = sub->link_next)
{
asection *o;
if (! is_arm_elf (sub))
continue;
elf_shdrp = elf_elfsections (sub);
for (o = sub->sections; o != NULL; o = o->next)
{
Elf_Internal_Shdr *hdr;
hdr = &elf_section_data (o)->this_hdr;
if (hdr->sh_type == SHT_ARM_EXIDX
&& hdr->sh_link
&& hdr->sh_link < elf_numsections (sub)
&& !o->gc_mark
&& elf_shdrp[hdr->sh_link]->bfd_section->gc_mark)
{
again = TRUE;
if (!_bfd_elf_gc_mark (info, o, gc_mark_hook))
return FALSE;
}
}
}
}
return TRUE;
}
static bfd_boolean
elf32_arm_is_target_special_symbol (bfd * abfd ATTRIBUTE_UNUSED, asymbol * sym)
{
return bfd_is_arm_special_symbol_name (sym->name,
BFD_ARM_SPECIAL_SYM_TYPE_ANY);
}
ARM mapping symbols are ignored when looking for function names
and STT_ARM_TFUNC is considered to a function type. */
static bfd_boolean
arm_elf_find_function (bfd * abfd ATTRIBUTE_UNUSED,
asection * section,
asymbol ** symbols,
bfd_vma offset,
const char ** filename_ptr,
const char ** functionname_ptr)
{
const char * filename = NULL;
asymbol * func = NULL;
bfd_vma low_func = 0;
asymbol ** p;
for (p = symbols; *p != NULL; p++)
{
elf_symbol_type *q;
q = (elf_symbol_type *) *p;
switch (ELF_ST_TYPE (q->internal_elf_sym.st_info))
{
default:
break;
case STT_FILE:
filename = bfd_asymbol_name (&q->symbol);
break;
case STT_FUNC:
case STT_ARM_TFUNC:
case STT_NOTYPE:
if ((q->symbol.flags & BSF_LOCAL)
&& bfd_is_arm_special_symbol_name (q->symbol.name,
BFD_ARM_SPECIAL_SYM_TYPE_ANY))
continue;
if (bfd_get_section (&q->symbol) == section
&& q->symbol.value >= low_func
&& q->symbol.value <= offset)
{
func = (asymbol *) q;
low_func = q->symbol.value;
}
break;
}
}
if (func == NULL)
return FALSE;
if (filename_ptr)
*filename_ptr = filename;
if (functionname_ptr)
*functionname_ptr = bfd_asymbol_name (func);
return TRUE;
}
reporting. This code is a duplicate of the code in elf.c, except
that it uses arm_elf_find_function. */
static bfd_boolean
elf32_arm_find_nearest_line (bfd * abfd,
asection * section,
asymbol ** symbols,
bfd_vma offset,
const char ** filename_ptr,
const char ** functionname_ptr,
unsigned int * line_ptr)
{
bfd_boolean found = FALSE;
if (_bfd_dwarf2_find_nearest_line (abfd, section, symbols, offset,
filename_ptr, functionname_ptr,
line_ptr, 0,
& elf_tdata (abfd)->dwarf2_find_line_info))
{
if (!*functionname_ptr)
arm_elf_find_function (abfd, section, symbols, offset,
*filename_ptr ? NULL : filename_ptr,
functionname_ptr);
return TRUE;
}
if (! _bfd_stab_section_find_nearest_line (abfd, symbols, section, offset,
& found, filename_ptr,
functionname_ptr, line_ptr,
& elf_tdata (abfd)->line_info))
return FALSE;
if (found && (*functionname_ptr || *line_ptr))
return TRUE;
if (symbols == NULL)
return FALSE;
if (! arm_elf_find_function (abfd, section, symbols, offset,
filename_ptr, functionname_ptr))
return FALSE;
*line_ptr = 0;
return TRUE;
}
static bfd_boolean
elf32_arm_find_inliner_info (bfd * abfd,
const char ** filename_ptr,
const char ** functionname_ptr,
unsigned int * line_ptr)
{
bfd_boolean found;
found = _bfd_dwarf2_find_inliner_info (abfd, filename_ptr,
functionname_ptr, line_ptr,
& elf_tdata (abfd)->dwarf2_find_line_info);
return found;
}
regular object. The current definition is in some section of the
dynamic object, but we're not including those sections. We have to
change the definition to something the rest of the link can
understand. */
static bfd_boolean
elf32_arm_adjust_dynamic_symbol (struct bfd_link_info * info,
struct elf_link_hash_entry * h)
{
bfd * dynobj;
asection * s;
struct elf32_arm_link_hash_entry * eh;
struct elf32_arm_link_hash_table *globals;
globals = elf32_arm_hash_table (info);
dynobj = elf_hash_table (info)->dynobj;
BFD_ASSERT (dynobj != NULL
&& (h->needs_plt
|| h->u.weakdef != NULL
|| (h->def_dynamic
&& h->ref_regular
&& !h->def_regular)));
eh = (struct elf32_arm_link_hash_entry *) h;
will fill in the contents of the procedure linkage table later,
when we know the address of the .got section. */
if (h->type == STT_FUNC || h->type == STT_ARM_TFUNC
|| h->needs_plt)
{
if (h->plt.refcount <= 0
|| SYMBOL_CALLS_LOCAL (info, h)
|| (ELF_ST_VISIBILITY (h->other) != STV_DEFAULT
&& h->root.type == bfd_link_hash_undefweak))
{
file, but the symbol was never referred to by a dynamic
object, or if all references were garbage collected. In
such a case, we don't actually need to build a procedure
linkage table, and we can just do a PC24 reloc instead. */
h->plt.offset = (bfd_vma) -1;
eh->plt_thumb_refcount = 0;
eh->plt_maybe_thumb_refcount = 0;
h->needs_plt = 0;
}
return TRUE;
}
else
{
needed for an R_ARM_PC24 or similar reloc to a non-function sym
in check_relocs. We can't decide accurately between function
and non-function syms in check-relocs; Objects loaded later in
the link may change h->type. So fix it now. */
h->plt.offset = (bfd_vma) -1;
eh->plt_thumb_refcount = 0;
eh->plt_maybe_thumb_refcount = 0;
}
processor independent code will have arranged for us to see the
real definition first, and we can just use the same value. */
if (h->u.weakdef != NULL)
{
BFD_ASSERT (h->u.weakdef->root.type == bfd_link_hash_defined
|| h->u.weakdef->root.type == bfd_link_hash_defweak);
h->root.u.def.section = h->u.weakdef->root.u.def.section;
h->root.u.def.value = h->u.weakdef->root.u.def.value;
return TRUE;
}
relocation. */
if (!h->non_got_ref)
return TRUE;
is not a function. */
only references to the symbol are via the global offset table.
For such cases we need not do anything here; the relocations will
be handled correctly by relocate_section. Relocatable executables
can reference data in shared objects directly, so we don't need to
do anything here. */
if (info->shared || globals->root.is_relocatable_executable)
return TRUE;
if (h->size == 0)
{
(*_bfd_error_handler) (_("dynamic variable `%s' is zero size"),
h->root.root.string);
return TRUE;
}
become part of the .bss section of the executable. There will be
an entry for this symbol in the .dynsym section. The dynamic
object will contain position independent code, so all references
from the dynamic object to this symbol will go through the global
offset table. The dynamic linker will use the .dynsym entry to
determine the address it must put in the global offset table, so
both the dynamic object and the regular object will refer to the
same memory location for the variable. */
s = bfd_get_section_by_name (dynobj, ".dynbss");
BFD_ASSERT (s != NULL);
copy the initial value out of the dynamic object and into the
runtime process image. We need to remember the offset into the
.rel(a).bss section we are going to use. */
if ((h->root.u.def.section->flags & SEC_ALLOC) != 0)
{
asection *srel;
srel = bfd_get_section_by_name (dynobj, RELOC_SECTION (globals, ".bss"));
BFD_ASSERT (srel != NULL);
srel->size += RELOC_SIZE (globals);
h->needs_copy = 1;
}
return _bfd_elf_adjust_dynamic_copy (h, s);
}
dynamic relocs. */
static bfd_boolean
allocate_dynrelocs (struct elf_link_hash_entry *h, void * inf)
{
struct bfd_link_info *info;
struct elf32_arm_link_hash_table *htab;
struct elf32_arm_link_hash_entry *eh;
struct elf32_arm_relocs_copied *p;
bfd_signed_vma thumb_refs;
eh = (struct elf32_arm_link_hash_entry *) h;
if (h->root.type == bfd_link_hash_indirect)
return TRUE;
if (h->root.type == bfd_link_hash_warning)
entry in the hash table, thus we never get to see the real
symbol in a hash traversal. So look at it now. */
h = (struct elf_link_hash_entry *) h->root.u.i.link;
info = (struct bfd_link_info *) inf;
htab = elf32_arm_hash_table (info);
if (htab->root.dynamic_sections_created
&& h->plt.refcount > 0)
{
Undefined weak syms won't yet be marked as dynamic. */
if (h->dynindx == -1
&& !h->forced_local)
{
if (! bfd_elf_link_record_dynamic_symbol (info, h))
return FALSE;
}
if (info->shared
|| WILL_CALL_FINISH_DYNAMIC_SYMBOL (1, 0, h))
{
asection *s = htab->splt;
first entry. */
if (s->size == 0)
s->size += htab->plt_header_size;
h->plt.offset = s->size;
for it. */
thumb_refs = eh->plt_thumb_refcount;
if (!htab->use_blx)
thumb_refs += eh->plt_maybe_thumb_refcount;
if (thumb_refs > 0)
{
h->plt.offset += PLT_THUMB_STUB_SIZE;
s->size += PLT_THUMB_STUB_SIZE;
}
not generating a shared library, then set the symbol to this
location in the .plt. This is required to make function
pointers compare as equal between the normal executable and
the shared library. */
if (! info->shared
&& !h->def_regular)
{
h->root.u.def.section = s;
h->root.u.def.value = h->plt.offset;
it is the target of an ABS32 relocation, which will
point to the PLT entry. */
if (ELF_ST_TYPE (h->type) == STT_ARM_TFUNC)
h->type = ELF_ST_INFO (ELF_ST_BIND (h->type), STT_FUNC);
}
s->size += htab->plt_entry_size;
if (!htab->symbian_p)
{
will be placed in the .got section by the linker script. */
eh->plt_got_offset = htab->sgotplt->size;
htab->sgotplt->size += 4;
}
htab->srelplt->size += RELOC_SIZE (htab);
each PLT entry. They go in a separate relocation section,
which is processed by the kernel loader. */
if (htab->vxworks_p && !info->shared)
{
an R_ARM_32 relocation for _GLOBAL_OFFSET_TABLE_. */
if (h->plt.offset == htab->plt_header_size)
htab->srelplt2->size += RELOC_SIZE (htab);
PLT entry: an R_ARM_32 relocation for the GOT entry,
and an R_ARM_32 relocation for the PLT entry. */
htab->srelplt2->size += RELOC_SIZE (htab) * 2;
}
}
else
{
h->plt.offset = (bfd_vma) -1;
h->needs_plt = 0;
}
}
else
{
h->plt.offset = (bfd_vma) -1;
h->needs_plt = 0;
}
if (h->got.refcount > 0)
{
asection *s;
bfd_boolean dyn;
int tls_type = elf32_arm_hash_entry (h)->tls_type;
int indx;
Undefined weak syms won't yet be marked as dynamic. */
if (h->dynindx == -1
&& !h->forced_local)
{
if (! bfd_elf_link_record_dynamic_symbol (info, h))
return FALSE;
}
if (!htab->symbian_p)
{
s = htab->sgot;
h->got.offset = s->size;
if (tls_type == GOT_UNKNOWN)
abort ();
if (tls_type == GOT_NORMAL)
s->size += 4;
else
{
if (tls_type & GOT_TLS_GD)
s->size += 8;
if (tls_type & GOT_TLS_IE)
s->size += 4;
}
dyn = htab->root.dynamic_sections_created;
indx = 0;
if (WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, info->shared, h)
&& (!info->shared
|| !SYMBOL_REFERENCES_LOCAL (info, h)))
indx = h->dynindx;
if (tls_type != GOT_NORMAL
&& (info->shared || indx != 0)
&& (ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
|| h->root.type != bfd_link_hash_undefweak))
{
if (tls_type & GOT_TLS_IE)
htab->srelgot->size += RELOC_SIZE (htab);
if (tls_type & GOT_TLS_GD)
htab->srelgot->size += RELOC_SIZE (htab);
if ((tls_type & GOT_TLS_GD) && indx != 0)
htab->srelgot->size += RELOC_SIZE (htab);
}
else if ((ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
|| h->root.type != bfd_link_hash_undefweak)
&& (info->shared
|| WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, 0, h)))
htab->srelgot->size += RELOC_SIZE (htab);
}
}
else
h->got.offset = (bfd_vma) -1;
if (!htab->use_blx && h->dynindx != -1
&& h->def_regular
&& ELF_ST_TYPE (h->type) == STT_ARM_TFUNC
&& ELF_ST_VISIBILITY (h->other) == STV_DEFAULT)
{
struct elf_link_hash_entry * th;
struct bfd_link_hash_entry * bh;
struct elf_link_hash_entry * myh;
char name[1024];
asection *s;
bh = NULL;
s = h->root.u.def.section;
sprintf (name, "__real_%s", h->root.root.string);
_bfd_generic_link_add_one_symbol (info, s->owner,
name, BSF_GLOBAL, s,
h->root.u.def.value,
NULL, TRUE, FALSE, &bh);
myh = (struct elf_link_hash_entry *) bh;
myh->type = ELF_ST_INFO (STB_LOCAL, STT_ARM_TFUNC);
myh->forced_local = 1;
eh->export_glue = myh;
th = record_arm_to_thumb_glue (info, h);
h->type = ELF_ST_INFO (ELF_ST_BIND (h->type), STT_FUNC);
h->root.u.def.section = th->root.u.def.section;
h->root.u.def.value = th->root.u.def.value & ~1;
}
if (eh->relocs_copied == NULL)
return TRUE;
dynamic pc-relative relocs against symbols which turn out to be
defined in regular objects. For the normal shared case, discard
space for pc-relative relocs that have become local due to symbol
visibility changes. */
if (info->shared || htab->root.is_relocatable_executable)
{
R_ARM_REL32_NOI, which will appear on something like
".long foo - .". We want calls to protected symbols to resolve
directly to the function rather than going via the plt. If people
want function pointer comparisons to work as expected then they
should avoid writing assembly like ".long foo - .". */
if (SYMBOL_CALLS_LOCAL (info, h))
{
struct elf32_arm_relocs_copied **pp;
for (pp = &eh->relocs_copied; (p = *pp) != NULL; )
{
p->count -= p->pc_count;
p->pc_count = 0;
if (p->count == 0)
*pp = p->next;
else
pp = &p->next;
}
}
if (elf32_arm_hash_table (info)->vxworks_p)
{
struct elf32_arm_relocs_copied **pp;
for (pp = &eh->relocs_copied; (p = *pp) != NULL; )
{
if (strcmp (p->section->output_section->name, ".tls_vars") == 0)
*pp = p->next;
else
pp = &p->next;
}
}
visibility. */
if (eh->relocs_copied != NULL
&& h->root.type == bfd_link_hash_undefweak)
{
if (ELF_ST_VISIBILITY (h->other) != STV_DEFAULT)
eh->relocs_copied = NULL;
symbol in PIEs. */
else if (h->dynindx == -1
&& !h->forced_local)
{
if (! bfd_elf_link_record_dynamic_symbol (info, h))
return FALSE;
}
}
else if (htab->root.is_relocatable_executable && h->dynindx == -1
&& h->root.type == bfd_link_hash_new)
{
against them. For normal symbols we output a relocation
against the section that contains them. */
if (! bfd_elf_link_record_dynamic_symbol (info, h))
return FALSE;
}
}
else
{
symbols which turn out to need copy relocs or are not
dynamic. */
if (!h->non_got_ref
&& ((h->def_dynamic
&& !h->def_regular)
|| (htab->root.dynamic_sections_created
&& (h->root.type == bfd_link_hash_undefweak
|| h->root.type == bfd_link_hash_undefined))))
{
Undefined weak syms won't yet be marked as dynamic. */
if (h->dynindx == -1
&& !h->forced_local)
{
if (! bfd_elf_link_record_dynamic_symbol (info, h))
return FALSE;
}
relocs. */
if (h->dynindx != -1)
goto keep;
}
eh->relocs_copied = NULL;
keep: ;
}
for (p = eh->relocs_copied; p != NULL; p = p->next)
{
asection *sreloc = elf_section_data (p->section)->sreloc;
sreloc->size += p->count * RELOC_SIZE (htab);
}
return TRUE;
}
static bfd_boolean
elf32_arm_readonly_dynrelocs (struct elf_link_hash_entry * h, void * inf)
{
struct elf32_arm_link_hash_entry * eh;
struct elf32_arm_relocs_copied * p;
if (h->root.type == bfd_link_hash_warning)
h = (struct elf_link_hash_entry *) h->root.u.i.link;
eh = (struct elf32_arm_link_hash_entry *) h;
for (p = eh->relocs_copied; p != NULL; p = p->next)
{
asection *s = p->section;
if (s != NULL && (s->flags & SEC_READONLY) != 0)
{
struct bfd_link_info *info = (struct bfd_link_info *) inf;
info->flags |= DF_TEXTREL;
return FALSE;
}
}
return TRUE;
}
void
bfd_elf32_arm_set_byteswap_code (struct bfd_link_info *info,
int byteswap_code)
{
struct elf32_arm_link_hash_table *globals;
globals = elf32_arm_hash_table (info);
globals->byteswap_code = byteswap_code;
}
static bfd_boolean
elf32_arm_size_dynamic_sections (bfd * output_bfd ATTRIBUTE_UNUSED,
struct bfd_link_info * info)
{
bfd * dynobj;
asection * s;
bfd_boolean plt;
bfd_boolean relocs;
bfd *ibfd;
struct elf32_arm_link_hash_table *htab;
htab = elf32_arm_hash_table (info);
dynobj = elf_hash_table (info)->dynobj;
BFD_ASSERT (dynobj != NULL);
check_use_blx (htab);
if (elf_hash_table (info)->dynamic_sections_created)
{
if (info->executable)
{
s = bfd_get_section_by_name (dynobj, ".interp");
BFD_ASSERT (s != NULL);
s->size = sizeof ELF_DYNAMIC_INTERPRETER;
s->contents = (unsigned char *) ELF_DYNAMIC_INTERPRETER;
}
}
relocs. */
for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
{
bfd_signed_vma *local_got;
bfd_signed_vma *end_local_got;
char *local_tls_type;
bfd_size_type locsymcount;
Elf_Internal_Shdr *symtab_hdr;
asection *srel;
bfd_boolean is_vxworks = elf32_arm_hash_table (info)->vxworks_p;
if (! is_arm_elf (ibfd))
continue;
for (s = ibfd->sections; s != NULL; s = s->next)
{
struct elf32_arm_relocs_copied *p;
for (p = elf_section_data (s)->local_dynrel; p != NULL; p = p->next)
{
if (!bfd_is_abs_section (p->section)
&& bfd_is_abs_section (p->section->output_section))
{
it is a copy of a linkonce section or due to
linker script /DISCARD/, so we'll be discarding
the relocs too. */
}
else if (is_vxworks
&& strcmp (p->section->output_section->name,
".tls_vars") == 0)
{
handled specially by the loader. */
}
else if (p->count != 0)
{
srel = elf_section_data (p->section)->sreloc;
srel->size += p->count * RELOC_SIZE (htab);
if ((p->section->output_section->flags & SEC_READONLY) != 0)
info->flags |= DF_TEXTREL;
}
}
}
local_got = elf_local_got_refcounts (ibfd);
if (!local_got)
continue;
symtab_hdr = & elf_symtab_hdr (ibfd);
locsymcount = symtab_hdr->sh_info;
end_local_got = local_got + locsymcount;
local_tls_type = elf32_arm_local_got_tls_type (ibfd);
s = htab->sgot;
srel = htab->srelgot;
for (; local_got < end_local_got; ++local_got, ++local_tls_type)
{
if (*local_got > 0)
{
*local_got = s->size;
if (*local_tls_type & GOT_TLS_GD)
s->size += 8;
if (*local_tls_type & GOT_TLS_IE)
s->size += 4;
if (*local_tls_type == GOT_NORMAL)
s->size += 4;
if (info->shared || *local_tls_type == GOT_TLS_GD)
srel->size += RELOC_SIZE (htab);
}
else
*local_got = (bfd_vma) -1;
}
}
if (htab->tls_ldm_got.refcount > 0)
{
for R_ARM_TLS_LDM32 relocations. */
htab->tls_ldm_got.offset = htab->sgot->size;
htab->sgot->size += 8;
if (info->shared)
htab->srelgot->size += RELOC_SIZE (htab);
}
else
htab->tls_ldm_got.offset = -1;
sym dynamic relocs. */
elf_link_hash_traverse (& htab->root, allocate_dynrelocs, info);
for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
{
if (! is_arm_elf (ibfd))
continue;
bfd_elf32_arm_init_maps (ibfd);
if (!bfd_elf32_arm_process_before_allocation (ibfd, info)
|| !bfd_elf32_arm_vfp11_erratum_scan (ibfd, info))
_bfd_error_handler (_("Errors encountered processing file %s"),
ibfd->filename);
}
determined the sizes of the various dynamic sections. Allocate
memory for them. */
plt = FALSE;
relocs = FALSE;
for (s = dynobj->sections; s != NULL; s = s->next)
{
const char * name;
if ((s->flags & SEC_LINKER_CREATED) == 0)
continue;
of the dynobj section names depend upon the input files. */
name = bfd_get_section_name (dynobj, s);
if (strcmp (name, ".plt") == 0)
{
plt = s->size != 0;
}
else if (CONST_STRNEQ (name, ".rel"))
{
if (s->size != 0)
{
than .rel(a).plt and .rela.plt.unloaded. */
if (s != htab->srelplt && s != htab->srelplt2)
relocs = TRUE;
to copy relocs into the output file. */
s->reloc_count = 0;
}
}
else if (! CONST_STRNEQ (name, ".got")
&& strcmp (name, ".dynbss") != 0)
{
continue;
}
if (s->size == 0)
{
output file. This is mostly to handle .rel(a).bss and
.rel(a).plt. We must create both sections in
create_dynamic_sections, because they must be created
before the linker maps input sections to output
sections. The linker does that before
adjust_dynamic_symbol is called, and it is that
function which decides whether anything needs to go
into these sections. */
s->flags |= SEC_EXCLUDE;
continue;
}
if ((s->flags & SEC_HAS_CONTENTS) == 0)
continue;
s->contents = bfd_zalloc (dynobj, s->size);
if (s->contents == NULL)
return FALSE;
}
if (elf_hash_table (info)->dynamic_sections_created)
{
values later, in elf32_arm_finish_dynamic_sections, but we
must add the entries now so that we get the correct size for
the .dynamic section. The DT_DEBUG entry is filled in by the
dynamic linker and used by the debugger. */
#define add_dynamic_entry(TAG, VAL) \
_bfd_elf_add_dynamic_entry (info, TAG, VAL)
if (info->executable)
{
if (!add_dynamic_entry (DT_DEBUG, 0))
return FALSE;
}
if (plt)
{
if ( !add_dynamic_entry (DT_PLTGOT, 0)
|| !add_dynamic_entry (DT_PLTRELSZ, 0)
|| !add_dynamic_entry (DT_PLTREL,
htab->use_rel ? DT_REL : DT_RELA)
|| !add_dynamic_entry (DT_JMPREL, 0))
return FALSE;
}
if (relocs)
{
if (htab->use_rel)
{
if (!add_dynamic_entry (DT_REL, 0)
|| !add_dynamic_entry (DT_RELSZ, 0)
|| !add_dynamic_entry (DT_RELENT, RELOC_SIZE (htab)))
return FALSE;
}
else
{
if (!add_dynamic_entry (DT_RELA, 0)
|| !add_dynamic_entry (DT_RELASZ, 0)
|| !add_dynamic_entry (DT_RELAENT, RELOC_SIZE (htab)))
return FALSE;
}
}
then we need a DT_TEXTREL entry. */
if ((info->flags & DF_TEXTREL) == 0)
elf_link_hash_traverse (& htab->root, elf32_arm_readonly_dynrelocs,
info);
if ((info->flags & DF_TEXTREL) != 0)
{
if (!add_dynamic_entry (DT_TEXTREL, 0))
return FALSE;
}
if (htab->vxworks_p
&& !elf_vxworks_add_dynamic_entries (output_bfd, info))
return FALSE;
}
#undef add_dynamic_entry
return TRUE;
}
dynamic sections here. */
static bfd_boolean
elf32_arm_finish_dynamic_symbol (bfd * output_bfd,
struct bfd_link_info * info,
struct elf_link_hash_entry * h,
Elf_Internal_Sym * sym)
{
bfd * dynobj;
struct elf32_arm_link_hash_table *htab;
struct elf32_arm_link_hash_entry *eh;
dynobj = elf_hash_table (info)->dynobj;
htab = elf32_arm_hash_table (info);
eh = (struct elf32_arm_link_hash_entry *) h;
if (h->plt.offset != (bfd_vma) -1)
{
asection * splt;
asection * srel;
bfd_byte *loc;
bfd_vma plt_index;
Elf_Internal_Rela rel;
it up. */
BFD_ASSERT (h->dynindx != -1);
splt = bfd_get_section_by_name (dynobj, ".plt");
srel = bfd_get_section_by_name (dynobj, RELOC_SECTION (htab, ".plt"));
BFD_ASSERT (splt != NULL && srel != NULL);
if (htab->symbian_p)
{
put_arm_insn (htab, output_bfd,
elf32_arm_symbian_plt_entry[0],
splt->contents + h->plt.offset);
bfd_put_32 (output_bfd,
elf32_arm_symbian_plt_entry[1],
splt->contents + h->plt.offset + 4);
rel.r_offset = (splt->output_section->vma
+ splt->output_offset
+ h->plt.offset + 4);
rel.r_info = ELF32_R_INFO (h->dynindx, R_ARM_GLOB_DAT);
corresponds to this symbol. This is the index of this symbol
in all the symbols for which we are making plt entries. The
first entry in the procedure linkage table is reserved. */
plt_index = ((h->plt.offset - htab->plt_header_size)
/ htab->plt_entry_size);
}
else
{
bfd_vma got_offset, got_address, plt_address;
bfd_vma got_displacement;
asection * sgot;
bfd_byte * ptr;
sgot = bfd_get_section_by_name (dynobj, ".got.plt");
BFD_ASSERT (sgot != NULL);
corresponds to this function. */
got_offset = eh->plt_got_offset;
corresponds to this symbol. This is the index of this symbol
in all the symbols for which we are making plt entries. The
first three entries in .got.plt are reserved; after that
symbols appear in the same order as in .plt. */
plt_index = (got_offset - 12) / 4;
got_address = (sgot->output_section->vma
+ sgot->output_offset
+ got_offset);
plt_address = (splt->output_section->vma
+ splt->output_offset
+ h->plt.offset);
ptr = htab->splt->contents + h->plt.offset;
if (htab->vxworks_p && info->shared)
{
unsigned int i;
bfd_vma val;
for (i = 0; i != htab->plt_entry_size / 4; i++, ptr += 4)
{
val = elf32_arm_vxworks_shared_plt_entry[i];
if (i == 2)
val |= got_address - sgot->output_section->vma;
if (i == 5)
val |= plt_index * RELOC_SIZE (htab);
if (i == 2 || i == 5)
bfd_put_32 (output_bfd, val, ptr);
else
put_arm_insn (htab, output_bfd, val, ptr);
}
}
else if (htab->vxworks_p)
{
unsigned int i;
bfd_vma val;
for (i = 0; i != htab->plt_entry_size / 4; i++, ptr += 4)
{
val = elf32_arm_vxworks_exec_plt_entry[i];
if (i == 2)
val |= got_address;
if (i == 4)
val |= 0xffffff & -((h->plt.offset + i * 4 + 8) >> 2);
if (i == 5)
val |= plt_index * RELOC_SIZE (htab);
if (i == 2 || i == 5)
bfd_put_32 (output_bfd, val, ptr);
else
put_arm_insn (htab, output_bfd, val, ptr);
}
loc = (htab->srelplt2->contents
+ (plt_index * 2 + 1) * RELOC_SIZE (htab));
referencing the GOT for this PLT entry. */
rel.r_offset = plt_address + 8;
rel.r_info = ELF32_R_INFO (htab->root.hgot->indx, R_ARM_ABS32);
rel.r_addend = got_offset;
SWAP_RELOC_OUT (htab) (output_bfd, &rel, loc);
loc += RELOC_SIZE (htab);
beginning of the PLT for this GOT entry. */
rel.r_offset = got_address;
rel.r_info = ELF32_R_INFO (htab->root.hplt->indx, R_ARM_ABS32);
rel.r_addend = 0;
SWAP_RELOC_OUT (htab) (output_bfd, &rel, loc);
}
else
{
bfd_signed_vma thumb_refs;
entry in the GOT. The eight-byte offset accounts for the
value produced by adding to pc in the first instruction
of the PLT stub. */
got_displacement = got_address - (plt_address + 8);
BFD_ASSERT ((got_displacement & 0xf0000000) == 0);
thumb_refs = eh->plt_thumb_refcount;
if (!htab->use_blx)
thumb_refs += eh->plt_maybe_thumb_refcount;
if (thumb_refs > 0)
{
put_thumb_insn (htab, output_bfd,
elf32_arm_plt_thumb_stub[0], ptr - 4);
put_thumb_insn (htab, output_bfd,
elf32_arm_plt_thumb_stub[1], ptr - 2);
}
put_arm_insn (htab, output_bfd,
elf32_arm_plt_entry[0]
| ((got_displacement & 0x0ff00000) >> 20),
ptr + 0);
put_arm_insn (htab, output_bfd,
elf32_arm_plt_entry[1]
| ((got_displacement & 0x000ff000) >> 12),
ptr+ 4);
put_arm_insn (htab, output_bfd,
elf32_arm_plt_entry[2]
| (got_displacement & 0x00000fff),
ptr + 8);
#ifdef FOUR_WORD_PLT
bfd_put_32 (output_bfd, elf32_arm_plt_entry[3], ptr + 12);
#endif
}
bfd_put_32 (output_bfd,
(splt->output_section->vma
+ splt->output_offset),
sgot->contents + got_offset);
rel.r_addend = 0;
rel.r_offset = got_address;
rel.r_info = ELF32_R_INFO (h->dynindx, R_ARM_JUMP_SLOT);
}
loc = srel->contents + plt_index * RELOC_SIZE (htab);
SWAP_RELOC_OUT (htab) (output_bfd, &rel, loc);
if (!h->def_regular)
{
the .plt section. Leave the value alone. */
sym->st_shndx = SHN_UNDEF;
Otherwise, the PLT entry would provide a definition for
the symbol even if the symbol wasn't defined anywhere,
and so the symbol would never be NULL. */
if (!h->ref_regular_nonweak)
sym->st_value = 0;
}
}
if (h->got.offset != (bfd_vma) -1
&& (elf32_arm_hash_entry (h)->tls_type & GOT_TLS_GD) == 0
&& (elf32_arm_hash_entry (h)->tls_type & GOT_TLS_IE) == 0)
{
asection * sgot;
asection * srel;
Elf_Internal_Rela rel;
bfd_byte *loc;
bfd_vma offset;
up. */
sgot = bfd_get_section_by_name (dynobj, ".got");
srel = bfd_get_section_by_name (dynobj, RELOC_SECTION (htab, ".got"));
BFD_ASSERT (sgot != NULL && srel != NULL);
offset = (h->got.offset & ~(bfd_vma) 1);
rel.r_addend = 0;
rel.r_offset = (sgot->output_section->vma
+ sgot->output_offset
+ offset);
symbol is defined locally or was forced to be local because
of a version file, we just want to emit a RELATIVE reloc.
The entry in the global offset table will already have been
initialized in the relocate_section function. */
if (info->shared
&& SYMBOL_REFERENCES_LOCAL (info, h))
{
BFD_ASSERT ((h->got.offset & 1) != 0);
rel.r_info = ELF32_R_INFO (0, R_ARM_RELATIVE);
if (!htab->use_rel)
{
rel.r_addend = bfd_get_32 (output_bfd, sgot->contents + offset);
bfd_put_32 (output_bfd, (bfd_vma) 0, sgot->contents + offset);
}
}
else
{
BFD_ASSERT ((h->got.offset & 1) == 0);
bfd_put_32 (output_bfd, (bfd_vma) 0, sgot->contents + offset);
rel.r_info = ELF32_R_INFO (h->dynindx, R_ARM_GLOB_DAT);
}
loc = srel->contents + srel->reloc_count++ * RELOC_SIZE (htab);
SWAP_RELOC_OUT (htab) (output_bfd, &rel, loc);
}
if (h->needs_copy)
{
asection * s;
Elf_Internal_Rela rel;
bfd_byte *loc;
BFD_ASSERT (h->dynindx != -1
&& (h->root.type == bfd_link_hash_defined
|| h->root.type == bfd_link_hash_defweak));
s = bfd_get_section_by_name (h->root.u.def.section->owner,
RELOC_SECTION (htab, ".bss"));
BFD_ASSERT (s != NULL);
rel.r_addend = 0;
rel.r_offset = (h->root.u.def.value
+ h->root.u.def.section->output_section->vma
+ h->root.u.def.section->output_offset);
rel.r_info = ELF32_R_INFO (h->dynindx, R_ARM_COPY);
loc = s->contents + s->reloc_count++ * RELOC_SIZE (htab);
SWAP_RELOC_OUT (htab) (output_bfd, &rel, loc);
}
the _GLOBAL_OFFSET_TABLE_ symbol is not absolute: it is relative
to the ".got" section. */
if (strcmp (h->root.root.string, "_DYNAMIC") == 0
|| (!htab->vxworks_p && h == htab->root.hgot))
sym->st_shndx = SHN_ABS;
return TRUE;
}
static bfd_boolean
elf32_arm_finish_dynamic_sections (bfd * output_bfd, struct bfd_link_info * info)
{
bfd * dynobj;
asection * sgot;
asection * sdyn;
dynobj = elf_hash_table (info)->dynobj;
sgot = bfd_get_section_by_name (dynobj, ".got.plt");
BFD_ASSERT (elf32_arm_hash_table (info)->symbian_p || sgot != NULL);
sdyn = bfd_get_section_by_name (dynobj, ".dynamic");
if (elf_hash_table (info)->dynamic_sections_created)
{
asection *splt;
Elf32_External_Dyn *dyncon, *dynconend;
struct elf32_arm_link_hash_table *htab;
htab = elf32_arm_hash_table (info);
splt = bfd_get_section_by_name (dynobj, ".plt");
BFD_ASSERT (splt != NULL && sdyn != NULL);
dyncon = (Elf32_External_Dyn *) sdyn->contents;
dynconend = (Elf32_External_Dyn *) (sdyn->contents + sdyn->size);
for (; dyncon < dynconend; dyncon++)
{
Elf_Internal_Dyn dyn;
const char * name;
asection * s;
bfd_elf32_swap_dyn_in (dynobj, dyncon, &dyn);
switch (dyn.d_tag)
{
unsigned int type;
default:
if (htab->vxworks_p
&& elf_vxworks_finish_dynamic_entry (output_bfd, &dyn))
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
break;
case DT_HASH:
name = ".hash";
goto get_vma_if_bpabi;
case DT_STRTAB:
name = ".dynstr";
goto get_vma_if_bpabi;
case DT_SYMTAB:
name = ".dynsym";
goto get_vma_if_bpabi;
case DT_VERSYM:
name = ".gnu.version";
goto get_vma_if_bpabi;
case DT_VERDEF:
name = ".gnu.version_d";
goto get_vma_if_bpabi;
case DT_VERNEED:
name = ".gnu.version_r";
goto get_vma_if_bpabi;
case DT_PLTGOT:
name = ".got";
goto get_vma;
case DT_JMPREL:
name = RELOC_SECTION (htab, ".plt");
get_vma:
s = bfd_get_section_by_name (output_bfd, name);
BFD_ASSERT (s != NULL);
if (!htab->symbian_p)
dyn.d_un.d_ptr = s->vma;
else
at the file offset, not the memory address, for the
convenience of the post linker. */
dyn.d_un.d_ptr = s->filepos;
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
break;
get_vma_if_bpabi:
if (htab->symbian_p)
goto get_vma;
break;
case DT_PLTRELSZ:
s = bfd_get_section_by_name (output_bfd,
RELOC_SECTION (htab, ".plt"));
BFD_ASSERT (s != NULL);
dyn.d_un.d_val = s->size;
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
break;
case DT_RELSZ:
case DT_RELASZ:
if (!htab->symbian_p)
{
procedure linkage table relocs (DT_JMPREL) should be
included in the overall relocs (DT_REL). This is
what Solaris does. However, UnixWare can not handle
that case. Therefore, we override the DT_RELSZ entry
here to make it not include the JMPREL relocs. Since
the linker script arranges for .rel(a).plt to follow all
other relocation sections, we don't have to worry
about changing the DT_REL entry. */
s = bfd_get_section_by_name (output_bfd,
RELOC_SECTION (htab, ".plt"));
if (s != NULL)
dyn.d_un.d_val -= s->size;
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
break;
}
case DT_REL:
case DT_RELA:
offset, not the VMA, of the first relocation
section. So, we use code similar to that in
elflink.c, but do not check for SHF_ALLOC on the
relcoation section, since relocations sections are
never allocated under the BPABI. The comments above
about Unixware notwithstanding, we include all of the
relocations here. */
if (htab->symbian_p)
{
unsigned int i;
type = ((dyn.d_tag == DT_REL || dyn.d_tag == DT_RELSZ)
? SHT_REL : SHT_RELA);
dyn.d_un.d_val = 0;
for (i = 1; i < elf_numsections (output_bfd); i++)
{
Elf_Internal_Shdr *hdr
= elf_elfsections (output_bfd)[i];
if (hdr->sh_type == type)
{
if (dyn.d_tag == DT_RELSZ
|| dyn.d_tag == DT_RELASZ)
dyn.d_un.d_val += hdr->sh_size;
else if ((ufile_ptr) hdr->sh_offset
<= dyn.d_un.d_val - 1)
dyn.d_un.d_val = hdr->sh_offset;
}
}
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
}
break;
corresponding function is Thumb. */
case DT_INIT:
name = info->init_function;
goto get_sym;
case DT_FINI:
name = info->fini_function;
get_sym:
then there is nothing to adjust. */
if (dyn.d_un.d_val != 0)
{
struct elf_link_hash_entry * eh;
eh = elf_link_hash_lookup (elf_hash_table (info), name,
FALSE, FALSE, TRUE);
if (eh != NULL
&& ELF_ST_TYPE (eh->type) == STT_ARM_TFUNC)
{
dyn.d_un.d_val |= 1;
bfd_elf32_swap_dyn_out (output_bfd, &dyn, dyncon);
}
}
break;
}
}
if (splt->size > 0 && elf32_arm_hash_table (info)->plt_header_size)
{
const bfd_vma *plt0_entry;
bfd_vma got_address, plt_address, got_displacement;
got_address = sgot->output_section->vma + sgot->output_offset;
plt_address = splt->output_section->vma + splt->output_offset;
if (htab->vxworks_p)
{
Therefore, we must emit relocations rather than simply
computing the values now. */
Elf_Internal_Rela rel;
plt0_entry = elf32_arm_vxworks_exec_plt0_entry;
put_arm_insn (htab, output_bfd, plt0_entry[0],
splt->contents + 0);
put_arm_insn (htab, output_bfd, plt0_entry[1],
splt->contents + 4);
put_arm_insn (htab, output_bfd, plt0_entry[2],
splt->contents + 8);
bfd_put_32 (output_bfd, got_address, splt->contents + 12);
rel.r_offset = plt_address + 12;
rel.r_info = ELF32_R_INFO (htab->root.hgot->indx, R_ARM_ABS32);
rel.r_addend = 0;
SWAP_RELOC_OUT (htab) (output_bfd, &rel,
htab->srelplt2->contents);
}
else
{
got_displacement = got_address - (plt_address + 16);
plt0_entry = elf32_arm_plt0_entry;
put_arm_insn (htab, output_bfd, plt0_entry[0],
splt->contents + 0);
put_arm_insn (htab, output_bfd, plt0_entry[1],
splt->contents + 4);
put_arm_insn (htab, output_bfd, plt0_entry[2],
splt->contents + 8);
put_arm_insn (htab, output_bfd, plt0_entry[3],
splt->contents + 12);
#ifdef FOUR_WORD_PLT
last word of the second entry. */
bfd_put_32 (output_bfd, got_displacement, splt->contents + 28);
#else
bfd_put_32 (output_bfd, got_displacement, splt->contents + 16);
#endif
}
}
really seem like the right value. */
if (splt->output_section->owner == output_bfd)
elf_section_data (splt->output_section)->this_hdr.sh_entsize = 4;
if (htab->vxworks_p && !info->shared && htab->splt->size > 0)
{
incorrect symbol indexes. */
int num_plts;
unsigned char *p;
num_plts = ((htab->splt->size - htab->plt_header_size)
/ htab->plt_entry_size);
p = htab->srelplt2->contents + RELOC_SIZE (htab);
for (; num_plts; num_plts--)
{
Elf_Internal_Rela rel;
SWAP_RELOC_IN (htab) (output_bfd, p, &rel);
rel.r_info = ELF32_R_INFO (htab->root.hgot->indx, R_ARM_ABS32);
SWAP_RELOC_OUT (htab) (output_bfd, &rel, p);
p += RELOC_SIZE (htab);
SWAP_RELOC_IN (htab) (output_bfd, p, &rel);
rel.r_info = ELF32_R_INFO (htab->root.hplt->indx, R_ARM_ABS32);
SWAP_RELOC_OUT (htab) (output_bfd, &rel, p);
p += RELOC_SIZE (htab);
}
}
}
if (sgot)
{
if (sgot->size > 0)
{
if (sdyn == NULL)
bfd_put_32 (output_bfd, (bfd_vma) 0, sgot->contents);
else
bfd_put_32 (output_bfd,
sdyn->output_section->vma + sdyn->output_offset,
sgot->contents);
bfd_put_32 (output_bfd, (bfd_vma) 0, sgot->contents + 4);
bfd_put_32 (output_bfd, (bfd_vma) 0, sgot->contents + 8);
}
elf_section_data (sgot->output_section)->this_hdr.sh_entsize = 4;
}
return TRUE;
}
static void
elf32_arm_post_process_headers (bfd * abfd, struct bfd_link_info * link_info ATTRIBUTE_UNUSED)
{
Elf_Internal_Ehdr * i_ehdrp;
struct elf32_arm_link_hash_table *globals;
i_ehdrp = elf_elfheader (abfd);
if (EF_ARM_EABI_VERSION (i_ehdrp->e_flags) == EF_ARM_EABI_UNKNOWN)
i_ehdrp->e_ident[EI_OSABI] = ELFOSABI_ARM;
else
i_ehdrp->e_ident[EI_OSABI] = 0;
i_ehdrp->e_ident[EI_ABIVERSION] = ARM_ELF_ABI_VERSION;
if (link_info)
{
globals = elf32_arm_hash_table (link_info);
if (globals->byteswap_code)
i_ehdrp->e_flags |= EF_ARM_BE8;
}
}
static enum elf_reloc_type_class
elf32_arm_reloc_type_class (const Elf_Internal_Rela *rela)
{
switch ((int) ELF32_R_TYPE (rela->r_info))
{
case R_ARM_RELATIVE:
return reloc_class_relative;
case R_ARM_JUMP_SLOT:
return reloc_class_plt;
case R_ARM_COPY:
return reloc_class_copy;
default:
return reloc_class_normal;
}
}
static bfd_boolean
elf32_arm_section_flags (flagword *flags, const Elf_Internal_Shdr *hdr)
{
if (hdr->sh_type == SHT_NOTE)
*flags |= SEC_LINK_ONCE | SEC_LINK_DUPLICATES_SAME_CONTENTS;
return TRUE;
}
static void
elf32_arm_final_write_processing (bfd *abfd, bfd_boolean linker ATTRIBUTE_UNUSED)
{
bfd_arm_update_notes (abfd, ARM_NOTE_SECTION);
}
static bfd_boolean
is_arm_elf_unwind_section_name (bfd * abfd ATTRIBUTE_UNUSED, const char * name)
{
return (CONST_STRNEQ (name, ELF_STRING_ARM_unwind)
|| CONST_STRNEQ (name, ELF_STRING_ARM_unwind_once));
}
the section name, which is a hack, but ought to work. */
static bfd_boolean
elf32_arm_fake_sections (bfd * abfd, Elf_Internal_Shdr * hdr, asection * sec)
{
const char * name;
name = bfd_get_section_name (abfd, sec);
if (is_arm_elf_unwind_section_name (abfd, name))
{
hdr->sh_type = SHT_ARM_EXIDX;
hdr->sh_flags |= SHF_LINK_ORDER;
}
return TRUE;
}
called when bfd_section_from_shdr finds a section with an unknown
type. */
static bfd_boolean
elf32_arm_section_from_shdr (bfd *abfd,
Elf_Internal_Shdr * hdr,
const char *name,
int shindex)
{
at the moment there isn't one. We just keep track of the
sections by their name, instead. Fortunately, the ABI gives
names for all the ARM specific sections, so we will probably get
away with this. */
switch (hdr->sh_type)
{
case SHT_ARM_EXIDX:
case SHT_ARM_PREEMPTMAP:
case SHT_ARM_ATTRIBUTES:
break;
default:
return FALSE;
}
if (! _bfd_elf_make_section_from_shdr (abfd, hdr, name, shindex))
return FALSE;
return TRUE;
}
of the next and prev fields in the asection structure. */
typedef struct section_list
{
asection * sec;
struct section_list * next;
struct section_list * prev;
}
section_list;
an _arm_elf_section_data structure has been allocated. This
is because it is possible for functions like elf32_arm_write_section
to be called on a section which has had an elf_data_structure
allocated for it (and so the used_by_bfd field is valid) but
for which the ARM extended version of this structure - the
_arm_elf_section_data structure - has not been allocated. */
static section_list * sections_with_arm_elf_section_data = NULL;
static void
record_section_with_arm_elf_section_data (asection * sec)
{
struct section_list * entry;
entry = bfd_malloc (sizeof (* entry));
if (entry == NULL)
return;
entry->sec = sec;
entry->next = sections_with_arm_elf_section_data;
entry->prev = NULL;
if (entry->next != NULL)
entry->next->prev = entry;
sections_with_arm_elf_section_data = entry;
}
static struct section_list *
find_arm_elf_section_entry (asection * sec)
{
struct section_list * entry;
static struct section_list * last_entry = NULL;
to the sections_with_arm_elf_section_data list in forward order and
then looked up here in backwards order. This makes a real difference
to the ld-srec/sec64k.exp linker test. */
entry = sections_with_arm_elf_section_data;
if (last_entry != NULL)
{
if (last_entry->sec == sec)
entry = last_entry;
else if (last_entry->next != NULL
&& last_entry->next->sec == sec)
entry = last_entry->next;
}
for (; entry; entry = entry->next)
if (entry->sec == sec)
break;
if (entry)
likely to want to locate next time. Also this way if we have been
called from unrecord_section_with_arm_elf_section_data() we will not
be caching a pointer that is about to be freed. */
last_entry = entry->prev;
return entry;
}
static _arm_elf_section_data *
get_arm_elf_section_data (asection * sec)
{
struct section_list * entry;
entry = find_arm_elf_section_entry (sec);
if (entry)
return elf32_arm_section_data (entry->sec);
else
return NULL;
}
static void
unrecord_section_with_arm_elf_section_data (asection * sec)
{
struct section_list * entry;
entry = find_arm_elf_section_entry (sec);
if (entry)
{
if (entry->prev != NULL)
entry->prev->next = entry->next;
if (entry->next != NULL)
entry->next->prev = entry->prev;
if (entry == sections_with_arm_elf_section_data)
sections_with_arm_elf_section_data = entry->next;
free (entry);
}
}
typedef struct
{
void *finfo;
struct bfd_link_info *info;
asection *sec;
int sec_shndx;
bfd_boolean (*func) (void *, const char *, Elf_Internal_Sym *,
asection *, struct elf_link_hash_entry *);
} output_arch_syminfo;
enum map_symbol_type
{
ARM_MAP_ARM,
ARM_MAP_THUMB,
ARM_MAP_DATA
};
static bfd_boolean
elf32_arm_output_map_sym (output_arch_syminfo *osi,
enum map_symbol_type type,
bfd_vma offset)
{
static const char *names[3] = {"$a", "$t", "$d"};
struct elf32_arm_link_hash_table *htab;
Elf_Internal_Sym sym;
htab = elf32_arm_hash_table (osi->info);
sym.st_value = osi->sec->output_section->vma
+ osi->sec->output_offset
+ offset;
sym.st_size = 0;
sym.st_other = 0;
sym.st_info = ELF_ST_INFO (STB_LOCAL, STT_NOTYPE);
sym.st_shndx = osi->sec_shndx;
if (!osi->func (osi->finfo, names[type], &sym, osi->sec, NULL))
return FALSE;
return TRUE;
}
static bfd_boolean
elf32_arm_output_plt_map (struct elf_link_hash_entry *h, void *inf)
{
output_arch_syminfo *osi = (output_arch_syminfo *) inf;
struct elf32_arm_link_hash_table *htab;
struct elf32_arm_link_hash_entry *eh;
bfd_vma addr;
htab = elf32_arm_hash_table (osi->info);
if (h->root.type == bfd_link_hash_indirect)
return TRUE;
if (h->root.type == bfd_link_hash_warning)
entry in the hash table, thus we never get to see the real
symbol in a hash traversal. So look at it now. */
h = (struct elf_link_hash_entry *) h->root.u.i.link;
if (h->plt.offset == (bfd_vma) -1)
return TRUE;
eh = (struct elf32_arm_link_hash_entry *) h;
addr = h->plt.offset;
if (htab->symbian_p)
{
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 4))
return FALSE;
}
else if (htab->vxworks_p)
{
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 8))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr + 12))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 20))
return FALSE;
}
else
{
bfd_signed_vma thumb_refs;
thumb_refs = eh->plt_thumb_refcount;
if (!htab->use_blx)
thumb_refs += eh->plt_maybe_thumb_refcount;
if (thumb_refs > 0)
{
if (!elf32_arm_output_map_sym (osi, ARM_MAP_THUMB, addr - 4))
return FALSE;
}
#ifdef FOUR_WORD_PLT
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 12))
return FALSE;
#else
so only need to output a mapping symbol for the first PLT entry and
entries with thumb thunks. */
if (thumb_refs > 0 || addr == 20)
{
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
}
#endif
}
return TRUE;
}
static bfd_boolean
elf32_arm_output_stub_sym (output_arch_syminfo *osi, const char *name,
bfd_vma offset, bfd_vma size)
{
struct elf32_arm_link_hash_table *htab;
Elf_Internal_Sym sym;
htab = elf32_arm_hash_table (osi->info);
sym.st_value = osi->sec->output_section->vma
+ osi->sec->output_offset
+ offset;
sym.st_size = size;
sym.st_other = 0;
sym.st_info = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
sym.st_shndx = osi->sec_shndx;
if (!osi->func (osi->finfo, name, &sym, osi->sec, NULL))
return FALSE;
return TRUE;
}
static bfd_boolean
arm_map_one_stub (struct bfd_hash_entry * gen_entry,
void * in_arg)
{
struct elf32_arm_stub_hash_entry *stub_entry;
struct bfd_link_info *info;
struct elf32_arm_link_hash_table *htab;
asection *stub_sec;
bfd_vma addr;
char *stub_name;
output_arch_syminfo *osi;
stub_entry = (struct elf32_arm_stub_hash_entry *) gen_entry;
osi = (output_arch_syminfo *) in_arg;
info = osi->info;
htab = elf32_arm_hash_table (info);
stub_sec = stub_entry->stub_sec;
processed. */
if (stub_sec != osi->sec)
return TRUE;
addr = (bfd_vma) stub_entry->stub_offset;
stub_name = stub_entry->output_name;
switch (stub_entry->stub_type)
{
case arm_stub_long_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr, 8))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 4))
return FALSE;
break;
case arm_thumb_v4t_stub_long_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr, 12))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 8))
return FALSE;
break;
case arm_thumb_thumb_stub_long_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr | 1, 16))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_THUMB, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 12))
return FALSE;
break;
case arm_thumb_arm_v4t_stub_long_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr | 1, 20))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_THUMB, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr + 8))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 16))
return FALSE;
break;
case arm_thumb_arm_v4t_stub_short_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr | 1, 8))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr + 4))
return FALSE;
break;
case arm_stub_pic_long_branch:
if (!elf32_arm_output_stub_sym (osi, stub_name, addr, 12))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_ARM, addr))
return FALSE;
if (!elf32_arm_output_map_sym (osi, ARM_MAP_DATA, addr + 8))
return FALSE;
break;
default:
BFD_FAIL ();
}
return TRUE;
}
static bfd_boolean
elf32_arm_output_arch_local_syms (bfd *output_bfd,
struct bfd_link_info *info,
void *finfo,
bfd_boolean (*func) (void *, const char *,
Elf_Internal_Sym *,
asection *,
struct elf_link_hash_entry *))
{
output_arch_syminfo osi;
struct elf32_arm_link_hash_table *htab;
bfd_vma offset;
bfd_size_type size;
htab = elf32_arm_hash_table (info);
check_use_blx (htab);
osi.finfo = finfo;
osi.info = info;
osi.func = func;
if (htab->arm_glue_size > 0)
{
osi.sec = bfd_get_section_by_name (htab->bfd_of_glue_owner,
ARM2THUMB_GLUE_SECTION_NAME);
osi.sec_shndx = _bfd_elf_section_from_bfd_section
(output_bfd, osi.sec->output_section);
if (info->shared || htab->root.is_relocatable_executable
|| htab->pic_veneer)
size = ARM2THUMB_PIC_GLUE_SIZE;
else if (htab->use_blx)
size = ARM2THUMB_V5_STATIC_GLUE_SIZE;
else
size = ARM2THUMB_STATIC_GLUE_SIZE;
for (offset = 0; offset < htab->arm_glue_size; offset += size)
{
elf32_arm_output_map_sym (&osi, ARM_MAP_ARM, offset);
elf32_arm_output_map_sym (&osi, ARM_MAP_DATA, offset + size - 4);
}
}
if (htab->thumb_glue_size > 0)
{
osi.sec = bfd_get_section_by_name (htab->bfd_of_glue_owner,
THUMB2ARM_GLUE_SECTION_NAME);
osi.sec_shndx = _bfd_elf_section_from_bfd_section
(output_bfd, osi.sec->output_section);
size = THUMB2ARM_GLUE_SIZE;
for (offset = 0; offset < htab->thumb_glue_size; offset += size)
{
elf32_arm_output_map_sym (&osi, ARM_MAP_THUMB, offset);
elf32_arm_output_map_sym (&osi, ARM_MAP_ARM, offset + 4);
}
}
if (htab->bx_glue_size > 0)
{
osi.sec = bfd_get_section_by_name (htab->bfd_of_glue_owner,
ARM_BX_GLUE_SECTION_NAME);
osi.sec_shndx = _bfd_elf_section_from_bfd_section
(output_bfd, osi.sec->output_section);
elf32_arm_output_map_sym (&osi, ARM_MAP_ARM, 0);
}
if (htab->stub_bfd && htab->stub_bfd->sections)
{
asection* stub_sec;
for (stub_sec = htab->stub_bfd->sections;
stub_sec != NULL;
stub_sec = stub_sec->next)
{
if (!strstr (stub_sec->name, STUB_SUFFIX))
continue;
osi.sec = stub_sec;
osi.sec_shndx = _bfd_elf_section_from_bfd_section
(output_bfd, osi.sec->output_section);
bfd_hash_traverse (&htab->stub_hash_table, arm_map_one_stub, &osi);
}
}
if (!htab->splt || htab->splt->size == 0)
return TRUE;
osi.sec_shndx = _bfd_elf_section_from_bfd_section (output_bfd,
htab->splt->output_section);
osi.sec = htab->splt;
plt header. */
if (htab->vxworks_p)
{
if (!info->shared)
{
if (!elf32_arm_output_map_sym (&osi, ARM_MAP_ARM, 0))
return FALSE;
if (!elf32_arm_output_map_sym (&osi, ARM_MAP_DATA, 12))
return FALSE;
}
}
else if (!htab->symbian_p)
{
if (!elf32_arm_output_map_sym (&osi, ARM_MAP_ARM, 0))
return FALSE;
#ifndef FOUR_WORD_PLT
if (!elf32_arm_output_map_sym (&osi, ARM_MAP_DATA, 16))
return FALSE;
#endif
}
elf_link_hash_traverse (&htab->root, elf32_arm_output_plt_map, (void *) &osi);
return TRUE;
}
static bfd_boolean
elf32_arm_new_section_hook (bfd *abfd, asection *sec)
{
if (!sec->used_by_bfd)
{
_arm_elf_section_data *sdata;
bfd_size_type amt = sizeof (*sdata);
sdata = bfd_zalloc (abfd, amt);
if (sdata == NULL)
return FALSE;
sec->used_by_bfd = sdata;
}
record_section_with_arm_elf_section_data (sec);
return _bfd_elf_new_section_hook (abfd, sec);
}
static int
elf32_arm_compare_mapping (const void * a, const void * b)
{
const elf32_arm_section_map *amap = (const elf32_arm_section_map *) a;
const elf32_arm_section_map *bmap = (const elf32_arm_section_map *) b;
if (amap->vma > bmap->vma)
return 1;
else if (amap->vma < bmap->vma)
return -1;
else if (amap->type > bmap->type)
multiple mapping symbols at the same address by sorting on type
after vma. */
return 1;
else if (amap->type < bmap->type)
return -1;
else
return 0;
}
written out as normal. */
static bfd_boolean
elf32_arm_write_section (bfd *output_bfd,
struct bfd_link_info *link_info,
asection *sec,
bfd_byte *contents)
{
int mapcount, errcount;
_arm_elf_section_data *arm_data;
struct elf32_arm_link_hash_table *globals = elf32_arm_hash_table (link_info);
elf32_arm_section_map *map;
elf32_vfp11_erratum_list *errnode;
bfd_vma ptr;
bfd_vma end;
bfd_vma offset = sec->output_section->vma + sec->output_offset;
bfd_byte tmp;
int i;
structure then we cannot record anything. */
arm_data = get_arm_elf_section_data (sec);
if (arm_data == NULL)
return FALSE;
mapcount = arm_data->mapcount;
map = arm_data->map;
errcount = arm_data->erratumcount;
if (errcount != 0)
{
unsigned int endianflip = bfd_big_endian (output_bfd) ? 3 : 0;
for (errnode = arm_data->erratumlist; errnode != 0;
errnode = errnode->next)
{
bfd_vma index = errnode->vma - offset;
switch (errnode->type)
{
case VFP11_ERRATUM_BRANCH_TO_ARM_VENEER:
{
bfd_vma branch_to_veneer;
ARM B instruction. */
unsigned int insn = (errnode->u.b.vfp_insn & 0xf0000000)
| 0x0a000000;
index -= 4;
branch_to_veneer = errnode->u.b.veneer->vma
- errnode->vma - 4;
if ((signed) branch_to_veneer < -(1 << 25)
|| (signed) branch_to_veneer >= (1 << 25))
(*_bfd_error_handler) (_("%B: error: VFP11 veneer out of "
"range"), output_bfd);
insn |= (branch_to_veneer >> 2) & 0xffffff;
contents[endianflip ^ index] = insn & 0xff;
contents[endianflip ^ (index + 1)] = (insn >> 8) & 0xff;
contents[endianflip ^ (index + 2)] = (insn >> 16) & 0xff;
contents[endianflip ^ (index + 3)] = (insn >> 24) & 0xff;
}
break;
case VFP11_ERRATUM_ARM_VENEER:
{
bfd_vma branch_from_veneer;
unsigned int insn;
branch_from_veneer = errnode->u.v.branch->vma
- errnode->vma - 12;
if ((signed) branch_from_veneer < -(1 << 25)
|| (signed) branch_from_veneer >= (1 << 25))
(*_bfd_error_handler) (_("%B: error: VFP11 veneer out of "
"range"), output_bfd);
insn = errnode->u.v.branch->u.b.vfp_insn;
contents[endianflip ^ index] = insn & 0xff;
contents[endianflip ^ (index + 1)] = (insn >> 8) & 0xff;
contents[endianflip ^ (index + 2)] = (insn >> 16) & 0xff;
contents[endianflip ^ (index + 3)] = (insn >> 24) & 0xff;
insn = 0xea000000 | ((branch_from_veneer >> 2) & 0xffffff);
contents[endianflip ^ (index + 4)] = insn & 0xff;
contents[endianflip ^ (index + 5)] = (insn >> 8) & 0xff;
contents[endianflip ^ (index + 6)] = (insn >> 16) & 0xff;
contents[endianflip ^ (index + 7)] = (insn >> 24) & 0xff;
}
break;
default:
abort ();
}
}
}
if (mapcount == 0)
return FALSE;
if (globals->byteswap_code)
{
qsort (map, mapcount, sizeof (* map), elf32_arm_compare_mapping);
ptr = map[0].vma;
for (i = 0; i < mapcount; i++)
{
if (i == mapcount - 1)
end = sec->size;
else
end = map[i + 1].vma;
switch (map[i].type)
{
case 'a':
while (ptr + 3 < end)
{
tmp = contents[ptr];
contents[ptr] = contents[ptr + 3];
contents[ptr + 3] = tmp;
tmp = contents[ptr + 1];
contents[ptr + 1] = contents[ptr + 2];
contents[ptr + 2] = tmp;
ptr += 4;
}
break;
case 't':
while (ptr + 1 < end)
{
tmp = contents[ptr];
contents[ptr] = contents[ptr + 1];
contents[ptr + 1] = tmp;
ptr += 2;
}
break;
case 'd':
break;
}
ptr = end;
}
}
free (map);
arm_data->mapcount = 0;
arm_data->mapsize = 0;
arm_data->map = NULL;
unrecord_section_with_arm_elf_section_data (sec);
return FALSE;
}
static void
unrecord_section_via_map_over_sections (bfd * abfd ATTRIBUTE_UNUSED,
asection * sec,
void * ignore ATTRIBUTE_UNUSED)
{
unrecord_section_with_arm_elf_section_data (sec);
}
static bfd_boolean
elf32_arm_close_and_cleanup (bfd * abfd)
{
if (abfd->sections)
bfd_map_over_sections (abfd,
unrecord_section_via_map_over_sections,
NULL);
return _bfd_elf_close_and_cleanup (abfd);
}
static bfd_boolean
elf32_arm_bfd_free_cached_info (bfd * abfd)
{
if (abfd->sections)
bfd_map_over_sections (abfd,
unrecord_section_via_map_over_sections,
NULL);
return _bfd_free_cached_info (abfd);
}
static void
elf32_arm_symbol_processing (bfd *abfd ATTRIBUTE_UNUSED,
asymbol *asym)
{
elf_symbol_type *elfsym = (elf_symbol_type *) asym;
if (ELF_ST_TYPE (elfsym->internal_elf_sym.st_info) == STT_ARM_TFUNC)
elfsym->symbol.flags |= BSF_FUNCTION;
}
static bfd_boolean
elf32_arm_swap_symbol_in (bfd * abfd,
const void *psrc,
const void *pshn,
Elf_Internal_Sym *dst)
{
if (!bfd_elf32_swap_symbol_in (abfd, psrc, pshn, dst))
return FALSE;
the address. Turn these into STT_ARM_TFUNC. */
if (ELF_ST_TYPE (dst->st_info) == STT_FUNC
&& (dst->st_value & 1))
{
dst->st_info = ELF_ST_INFO (ELF_ST_BIND (dst->st_info), STT_ARM_TFUNC);
dst->st_value &= ~(bfd_vma) 1;
}
return TRUE;
}
static void
elf32_arm_swap_symbol_out (bfd *abfd,
const Elf_Internal_Sym *src,
void *cdst,
void *shndx)
{
Elf_Internal_Sym newsym;
of the address set, as per the new EABI. We do this unconditionally
because objcopy does not set the elf header flags until after
it writes out the symbol table. */
if (ELF_ST_TYPE (src->st_info) == STT_ARM_TFUNC)
{
newsym = *src;
newsym.st_info = ELF_ST_INFO (ELF_ST_BIND (src->st_info), STT_FUNC);
if (newsym.st_shndx != SHN_UNDEF)
{
linker will simulate the work of dynamic linker of resolving
symbols and will carry over the thumbness of found symbols to
the output symbol table. It's not clear how it happens, but
the thumbness of undefined symbols can well be different at
runtime, and writing '1' for them will be confusing for users
and possibly for dynamic linker itself.
*/
newsym.st_value |= 1;
}
src = &newsym;
}
bfd_elf32_swap_symbol_out (abfd, src, cdst, shndx);
}
static bfd_boolean
elf32_arm_modify_segment_map (bfd *abfd,
struct bfd_link_info *info ATTRIBUTE_UNUSED)
{
struct elf_segment_map *m;
asection *sec;
sec = bfd_get_section_by_name (abfd, ".ARM.exidx");
if (sec != NULL && (sec->flags & SEC_LOAD) != 0)
{
want to add another one. This situation arises when running
"strip"; the input binary already has the header. */
m = elf_tdata (abfd)->segment_map;
while (m && m->p_type != PT_ARM_EXIDX)
m = m->next;
if (!m)
{
m = bfd_zalloc (abfd, sizeof (struct elf_segment_map));
if (m == NULL)
return FALSE;
m->p_type = PT_ARM_EXIDX;
m->count = 1;
m->sections[0] = sec;
m->next = elf_tdata (abfd)->segment_map;
elf_tdata (abfd)->segment_map = m;
}
}
return TRUE;
}
static int
elf32_arm_additional_program_headers (bfd *abfd,
struct bfd_link_info *info ATTRIBUTE_UNUSED)
{
asection *sec;
sec = bfd_get_section_by_name (abfd, ".ARM.exidx");
if (sec != NULL && (sec->flags & SEC_LOAD) != 0)
return 1;
else
return 0;
}
static bfd_boolean
elf32_arm_is_function_type (unsigned int type)
{
return (type == STT_FUNC) || (type == STT_ARM_TFUNC);
}
const struct elf_size_info elf32_arm_size_info =
{
sizeof (Elf32_External_Ehdr),
sizeof (Elf32_External_Phdr),
sizeof (Elf32_External_Shdr),
sizeof (Elf32_External_Rel),
sizeof (Elf32_External_Rela),
sizeof (Elf32_External_Sym),
sizeof (Elf32_External_Dyn),
sizeof (Elf_External_Note),
4,
1,
32, 2,
ELFCLASS32, EV_CURRENT,
bfd_elf32_write_out_phdrs,
bfd_elf32_write_shdrs_and_ehdr,
bfd_elf32_checksum_contents,
bfd_elf32_write_relocs,
elf32_arm_swap_symbol_in,
elf32_arm_swap_symbol_out,
bfd_elf32_slurp_reloc_table,
bfd_elf32_slurp_symbol_table,
bfd_elf32_swap_dyn_in,
bfd_elf32_swap_dyn_out,
bfd_elf32_swap_reloc_in,
bfd_elf32_swap_reloc_out,
bfd_elf32_swap_reloca_in,
bfd_elf32_swap_reloca_out
};
#define ELF_ARCH bfd_arch_arm
#define ELF_MACHINE_CODE EM_ARM
#ifdef __QNXTARGET__
#define ELF_MAXPAGESIZE 0x1000
#else
#define ELF_MAXPAGESIZE 0x8000
#endif
#define ELF_MINPAGESIZE 0x1000
#define ELF_COMMONPAGESIZE 0x1000
#define bfd_elf32_mkobject elf32_arm_mkobject
#define bfd_elf32_bfd_copy_private_bfd_data elf32_arm_copy_private_bfd_data
#define bfd_elf32_bfd_merge_private_bfd_data elf32_arm_merge_private_bfd_data
#define bfd_elf32_bfd_set_private_flags elf32_arm_set_private_flags
#define bfd_elf32_bfd_print_private_bfd_data elf32_arm_print_private_bfd_data
#define bfd_elf32_bfd_link_hash_table_create elf32_arm_link_hash_table_create
#define bfd_elf32_bfd_link_hash_table_free elf32_arm_hash_table_free
#define bfd_elf32_bfd_reloc_type_lookup elf32_arm_reloc_type_lookup
#define bfd_elf32_bfd_reloc_name_lookup elf32_arm_reloc_name_lookup
#define bfd_elf32_find_nearest_line elf32_arm_find_nearest_line
#define bfd_elf32_find_inliner_info elf32_arm_find_inliner_info
#define bfd_elf32_new_section_hook elf32_arm_new_section_hook
#define bfd_elf32_bfd_is_target_special_symbol elf32_arm_is_target_special_symbol
#define bfd_elf32_close_and_cleanup elf32_arm_close_and_cleanup
#define bfd_elf32_bfd_free_cached_info elf32_arm_bfd_free_cached_info
#define elf_backend_get_symbol_type elf32_arm_get_symbol_type
#define elf_backend_gc_mark_hook elf32_arm_gc_mark_hook
#define elf_backend_gc_mark_extra_sections elf32_arm_gc_mark_extra_sections
#define elf_backend_gc_sweep_hook elf32_arm_gc_sweep_hook
#define elf_backend_check_relocs elf32_arm_check_relocs
#define elf_backend_relocate_section elf32_arm_relocate_section
#define elf_backend_write_section elf32_arm_write_section
#define elf_backend_adjust_dynamic_symbol elf32_arm_adjust_dynamic_symbol
#define elf_backend_create_dynamic_sections elf32_arm_create_dynamic_sections
#define elf_backend_finish_dynamic_symbol elf32_arm_finish_dynamic_symbol
#define elf_backend_finish_dynamic_sections elf32_arm_finish_dynamic_sections
#define elf_backend_size_dynamic_sections elf32_arm_size_dynamic_sections
#define elf_backend_init_index_section _bfd_elf_init_2_index_sections
#define elf_backend_post_process_headers elf32_arm_post_process_headers
#define elf_backend_reloc_type_class elf32_arm_reloc_type_class
#define elf_backend_object_p elf32_arm_object_p
#define elf_backend_section_flags elf32_arm_section_flags
#define elf_backend_fake_sections elf32_arm_fake_sections
#define elf_backend_section_from_shdr elf32_arm_section_from_shdr
#define elf_backend_final_write_processing elf32_arm_final_write_processing
#define elf_backend_copy_indirect_symbol elf32_arm_copy_indirect_symbol
#define elf_backend_symbol_processing elf32_arm_symbol_processing
#define elf_backend_size_info elf32_arm_size_info
#define elf_backend_modify_segment_map elf32_arm_modify_segment_map
#define elf_backend_additional_program_headers elf32_arm_additional_program_headers
#define elf_backend_output_arch_local_syms elf32_arm_output_arch_local_syms
#define elf_backend_begin_write_processing elf32_arm_begin_write_processing
#define elf_backend_is_function_type elf32_arm_is_function_type
#define elf_backend_can_refcount 1
#define elf_backend_can_gc_sections 1
#define elf_backend_plt_readonly 1
#define elf_backend_want_got_plt 1
#define elf_backend_want_plt_sym 0
#define elf_backend_may_use_rel_p 1
#define elf_backend_may_use_rela_p 0
#define elf_backend_default_use_rela_p 0
#define elf_backend_got_header_size 12
#undef elf_backend_obj_attrs_vendor
#define elf_backend_obj_attrs_vendor "aeabi"
#undef elf_backend_obj_attrs_section
#define elf_backend_obj_attrs_section ".ARM.attributes"
#undef elf_backend_obj_attrs_arg_type
#define elf_backend_obj_attrs_arg_type elf32_arm_obj_attrs_arg_type
#undef elf_backend_obj_attrs_section_type
#define elf_backend_obj_attrs_section_type SHT_ARM_ATTRIBUTES
#include "elf32-target.h"
#undef TARGET_LITTLE_SYM
#define TARGET_LITTLE_SYM bfd_elf32_littlearm_vxworks_vec
#undef TARGET_LITTLE_NAME
#define TARGET_LITTLE_NAME "elf32-littlearm-vxworks"
#undef TARGET_BIG_SYM
#define TARGET_BIG_SYM bfd_elf32_bigarm_vxworks_vec
#undef TARGET_BIG_NAME
#define TARGET_BIG_NAME "elf32-bigarm-vxworks"
appropriately for VxWorks. */
static struct bfd_link_hash_table *
elf32_arm_vxworks_link_hash_table_create (bfd *abfd)
{
struct bfd_link_hash_table *ret;
ret = elf32_arm_link_hash_table_create (abfd);
if (ret)
{
struct elf32_arm_link_hash_table *htab
= (struct elf32_arm_link_hash_table *) ret;
htab->use_rel = 0;
htab->vxworks_p = 1;
}
return ret;
}
static void
elf32_arm_vxworks_final_write_processing (bfd *abfd, bfd_boolean linker)
{
elf32_arm_final_write_processing (abfd, linker);
elf_vxworks_final_write_processing (abfd, linker);
}
#undef elf32_bed
#define elf32_bed elf32_arm_vxworks_bed
#undef bfd_elf32_bfd_link_hash_table_create
#define bfd_elf32_bfd_link_hash_table_create elf32_arm_vxworks_link_hash_table_create
#undef elf_backend_add_symbol_hook
#define elf_backend_add_symbol_hook elf_vxworks_add_symbol_hook
#undef elf_backend_final_write_processing
#define elf_backend_final_write_processing elf32_arm_vxworks_final_write_processing
#undef elf_backend_emit_relocs
#define elf_backend_emit_relocs elf_vxworks_emit_relocs
#undef elf_backend_may_use_rel_p
#define elf_backend_may_use_rel_p 0
#undef elf_backend_may_use_rela_p
#define elf_backend_may_use_rela_p 1
#undef elf_backend_default_use_rela_p
#define elf_backend_default_use_rela_p 1
#undef elf_backend_want_plt_sym
#define elf_backend_want_plt_sym 1
#undef ELF_MAXPAGESIZE
#define ELF_MAXPAGESIZE 0x1000
#include "elf32-target.h"
#undef TARGET_LITTLE_SYM
#define TARGET_LITTLE_SYM bfd_elf32_littlearm_symbian_vec
#undef TARGET_LITTLE_NAME
#define TARGET_LITTLE_NAME "elf32-littlearm-symbian"
#undef TARGET_BIG_SYM
#define TARGET_BIG_SYM bfd_elf32_bigarm_symbian_vec
#undef TARGET_BIG_NAME
#define TARGET_BIG_NAME "elf32-bigarm-symbian"
appropriately for Symbian OS. */
static struct bfd_link_hash_table *
elf32_arm_symbian_link_hash_table_create (bfd *abfd)
{
struct bfd_link_hash_table *ret;
ret = elf32_arm_link_hash_table_create (abfd);
if (ret)
{
struct elf32_arm_link_hash_table *htab
= (struct elf32_arm_link_hash_table *)ret;
htab->plt_header_size = 0;
htab->plt_entry_size = 4 * ARRAY_SIZE (elf32_arm_symbian_plt_entry);
htab->symbian_p = 1;
htab->use_blx = 1;
htab->root.is_relocatable_executable = 1;
}
return ret;
}
static const struct bfd_elf_special_section
elf32_arm_symbian_special_sections[] =
{
the loadable read-only segment. The post-linker may wish to
refer to these sections, but they are not part of the final
program image. */
{ STRING_COMMA_LEN (".dynamic"), 0, SHT_DYNAMIC, 0 },
{ STRING_COMMA_LEN (".dynstr"), 0, SHT_STRTAB, 0 },
{ STRING_COMMA_LEN (".dynsym"), 0, SHT_DYNSYM, 0 },
{ STRING_COMMA_LEN (".got"), 0, SHT_PROGBITS, 0 },
{ STRING_COMMA_LEN (".hash"), 0, SHT_HASH, 0 },
postlinker will arrange things so that no dynamic relocation is
required. */
{ STRING_COMMA_LEN (".init_array"), 0, SHT_INIT_ARRAY, SHF_ALLOC },
{ STRING_COMMA_LEN (".fini_array"), 0, SHT_FINI_ARRAY, SHF_ALLOC },
{ STRING_COMMA_LEN (".preinit_array"), 0, SHT_PREINIT_ARRAY, SHF_ALLOC },
{ NULL, 0, 0, 0, 0 }
};
static void
elf32_arm_symbian_begin_write_processing (bfd *abfd,
struct bfd_link_info *link_info)
{
processed by a postlinker first, into an OS-specific format. If
the D_PAGED bit is set on the file, BFD will align segments on
page boundaries, so that an OS can directly map the file. With
BPABI objects, that just results in wasted space. In addition,
because we clear the D_PAGED bit, map_sections_to_segments will
recognize that the program headers should not be mapped into any
loadable segment. */
abfd->flags &= ~D_PAGED;
elf32_arm_begin_write_processing (abfd, link_info);
}
static bfd_boolean
elf32_arm_symbian_modify_segment_map (bfd *abfd,
struct bfd_link_info *info)
{
struct elf_segment_map *m;
asection *dynsec;
segment. However, because the .dynamic section is not marked
with SEC_LOAD, the generic ELF code will not create such a
segment. */
dynsec = bfd_get_section_by_name (abfd, ".dynamic");
if (dynsec)
{
for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
if (m->p_type == PT_DYNAMIC)
break;
if (m == NULL)
{
m = _bfd_elf_make_dynamic_segment (abfd, dynsec);
m->next = elf_tdata (abfd)->segment_map;
elf_tdata (abfd)->segment_map = m;
}
}
return elf32_arm_modify_segment_map (abfd, info);
}
or (bfd_vma) -1 if it should not be included. */
static bfd_vma
elf32_arm_symbian_plt_sym_val (bfd_vma i, const asection *plt,
const arelent *rel ATTRIBUTE_UNUSED)
{
return plt->vma + 4 * ARRAY_SIZE (elf32_arm_symbian_plt_entry) * i;
}
#undef elf32_bed
#define elf32_bed elf32_arm_symbian_bed
will process them and then discard them. */
#undef ELF_DYNAMIC_SEC_FLAGS
#define ELF_DYNAMIC_SEC_FLAGS \
(SEC_HAS_CONTENTS | SEC_IN_MEMORY | SEC_LINKER_CREATED)
#undef elf_backend_add_symbol_hook
#undef elf_backend_emit_relocs
#undef bfd_elf32_bfd_link_hash_table_create
#define bfd_elf32_bfd_link_hash_table_create elf32_arm_symbian_link_hash_table_create
#undef elf_backend_special_sections
#define elf_backend_special_sections elf32_arm_symbian_special_sections
#undef elf_backend_begin_write_processing
#define elf_backend_begin_write_processing elf32_arm_symbian_begin_write_processing
#undef elf_backend_final_write_processing
#define elf_backend_final_write_processing elf32_arm_final_write_processing
#undef elf_backend_modify_segment_map
#define elf_backend_modify_segment_map elf32_arm_symbian_modify_segment_map
#undef elf_backend_got_header_size
#define elf_backend_got_header_size 0
#undef elf_backend_want_got_plt
#define elf_backend_want_got_plt 0
#undef elf_backend_plt_sym_val
#define elf_backend_plt_sym_val elf32_arm_symbian_plt_sym_val
#undef elf_backend_may_use_rel_p
#define elf_backend_may_use_rel_p 1
#undef elf_backend_may_use_rela_p
#define elf_backend_may_use_rela_p 0
#undef elf_backend_default_use_rela_p
#define elf_backend_default_use_rela_p 0
#undef elf_backend_want_plt_sym
#define elf_backend_want_plt_sym 0
#undef ELF_MAXPAGESIZE
#define ELF_MAXPAGESIZE 0x8000
#include "elf32-target.h"