#include "target.h"
#include "target-reloc.h"
#include "target-select.h"
+#include "tls.h"
namespace
{
// http://www.x86-64.org/documentation/abi.pdf
// TLS info comes from
// http://people.redhat.com/drepper/tls.pdf
-// http://ia32-abi.googlegroups.com/web/RFC-TLSDESC-x86.txt?gda=kWQJPEQAAACEfYQFX0dubPQ2NuO4whhjkR4HAp8tBMb_I0iuUeQslmG1qiJ7UbTIup-M2XPURDRiZJyPR4BqKR2agJ-5jfT5Ley2_-oiOJ4zLNAGCw24Bg
+// http://www.lsd.ic.unicamp.br/~oliva/writeups/TLS/RFC-TLSDESC-x86.txt
class Target_x86_64 : public Sized_target<64, false>
{
Target_x86_64()
: Sized_target<64, false>(&x86_64_info),
- got_(NULL), plt_(NULL), got_plt_(NULL), rel_dyn_(NULL),
+ got_(NULL), plt_(NULL), got_plt_(NULL), rela_dyn_(NULL),
copy_relocs_(NULL), dynbss_(NULL)
{ }
unsigned int sh_type,
const unsigned char* prelocs,
size_t reloc_count,
+ Output_section* output_section,
+ bool needs_special_offset_handling,
size_t local_symbol_count,
- const unsigned char* plocal_symbols,
- Symbol** global_symbols);
+ const unsigned char* plocal_symbols);
// Finalize the sections.
void
unsigned int sh_type,
const unsigned char* prelocs,
size_t reloc_count,
+ Output_section* output_section,
+ bool needs_special_offset_handling,
unsigned char* view,
elfcpp::Elf_types<64>::Elf_Addr view_address,
off_t view_size);
std::string
do_code_fill(off_t length);
+ // Return the size of the GOT section.
+ off_t
+ got_size()
+ {
+ gold_assert(this->got_ != NULL);
+ return this->got_->data_size();
+ }
+
private:
// The class which scans relocations.
struct Scan
unsigned int data_shndx,
const elfcpp::Rela<64, false>& reloc, unsigned int r_type,
Symbol* gsym);
+
+ static void
+ unsupported_reloc_local(Sized_relobj<64, false>*, unsigned int r_type);
+
+ static void
+ unsupported_reloc_global(Sized_relobj<64, false>*, unsigned int r_type,
+ Symbol*);
};
// The class which implements relocation.
if (this->skip_call_tls_get_addr_)
{
// FIXME: This needs to specify the location somehow.
- fprintf(stderr, _("%s: missing expected TLS relocation\n"),
- program_name);
- gold_exit(false);
+ gold_error(_("missing expected TLS relocation"));
}
}
const Symbol_value<64>*,
unsigned char*, elfcpp::Elf_types<64>::Elf_Addr, off_t);
- // Do a TLS Initial-Exec to Local-Exec transition.
- static inline void
- tls_ie_to_le(const Relocate_info<64, false>*, size_t relnum,
+ // Do a TLS General-Dynamic to Local-Exec transition.
+ inline void
+ tls_gd_to_le(const Relocate_info<64, false>*, size_t relnum,
Output_segment* tls_segment,
const elfcpp::Rela<64, false>&, unsigned int r_type,
elfcpp::Elf_types<64>::Elf_Addr value,
unsigned char* view,
off_t view_size);
- // Do a TLS Global-Dynamic to Local-Exec transition.
+ // Do a TLS Local-Dynamic to Local-Exec transition.
inline void
- tls_gd_to_le(const Relocate_info<64, false>*, size_t relnum,
+ tls_ld_to_le(const Relocate_info<64, false>*, size_t relnum,
Output_segment* tls_segment,
const elfcpp::Rela<64, false>&, unsigned int r_type,
elfcpp::Elf_types<64>::Elf_Addr value,
unsigned char* view,
off_t view_size);
- // Check the range for a TLS relocation.
- static inline void
- check_range(const Relocate_info<64, false>*, size_t relnum,
- const elfcpp::Rela<64, false>&, off_t, off_t);
-
- // Check the validity of a TLS relocation. This is like assert.
+ // Do a TLS Initial-Exec to Local-Exec transition.
static inline void
- check_tls(const Relocate_info<64, false>*, size_t relnum,
- const elfcpp::Rela<64, false>&, bool);
+ tls_ie_to_le(const Relocate_info<64, false>*, size_t relnum,
+ Output_segment* tls_segment,
+ const elfcpp::Rela<64, false>&, unsigned int r_type,
+ elfcpp::Elf_types<64>::Elf_Addr value,
+ unsigned char* view,
+ off_t view_size);
// This is set if we should skip the next reloc, which should be a
// PLT32 reloc against ___tls_get_addr.
// Adjust TLS relocation type based on the options and whether this
// is a local symbol.
- static unsigned int
+ static tls::Tls_optimization
optimize_tls_reloc(bool is_final, int r_type);
// Get the GOT section, creating it if necessary.
Output_data_got<64, false>*
got_section(Symbol_table*, Layout*);
+ // Get the GOT PLT section.
+ Output_data_space*
+ got_plt_section() const
+ {
+ gold_assert(this->got_plt_ != NULL);
+ return this->got_plt_;
+ }
+
// Create a PLT entry for a global symbol.
void
make_plt_entry(Symbol_table*, Layout*, Symbol*);
// Get the dynamic reloc section, creating it if necessary.
Reloc_section*
- rel_dyn_section(Layout*);
+ rela_dyn_section(Layout*);
+
+ // Return true if the symbol may need a COPY relocation.
+ // References from an executable object to non-function symbols
+ // defined in a dynamic object may need a COPY relocation.
+ bool
+ may_need_copy_reloc(Symbol* gsym)
+ {
+ return (!parameters->output_is_shared()
+ && gsym->is_from_dynobj()
+ && gsym->type() != elfcpp::STT_FUNC);
+ }
// Copy a relocation against a global symbol.
void
// The GOT PLT section.
Output_data_space* got_plt_;
// The dynamic reloc section.
- Reloc_section* rel_dyn_;
+ Reloc_section* rela_dyn_;
// Relocs saved to avoid a COPY reloc.
Copy_relocs<64, false>* copy_relocs_;
// Space for variables copied with a COPY reloc.
false, // has_make_symbol
false, // has_resolve
true, // has_code_fill
+ true, // is_default_stack_executable
"/lib/ld64.so.1", // program interpreter
- 0x400000, // text_segment_address
+ 0x400000, // default_text_segment_address
0x1000, // abi_pagesize
0x1000 // common_pagesize
};
// create another set of data in the .got section. Note that we
// always create a PLT if we create a GOT, although the PLT
// might be empty.
- // TODO(csilvers): do we really need an alignment of 8?
this->got_plt_ = new Output_data_space(8);
layout->add_output_section_data(".got", elfcpp::SHT_PROGBITS,
elfcpp::SHF_ALLOC | elfcpp::SHF_WRITE,
// Get the dynamic reloc section, creating it if necessary.
Target_x86_64::Reloc_section*
-Target_x86_64::rel_dyn_section(Layout* layout)
+Target_x86_64::rela_dyn_section(Layout* layout)
{
- if (this->rel_dyn_ == NULL)
+ if (this->rela_dyn_ == NULL)
{
gold_assert(layout != NULL);
- this->rel_dyn_ = new Reloc_section();
+ this->rela_dyn_ = new Reloc_section();
layout->add_output_section_data(".rela.dyn", elfcpp::SHT_RELA,
- elfcpp::SHF_ALLOC, this->rel_dyn_);
+ elfcpp::SHF_ALLOC, this->rela_dyn_);
}
- return this->rel_dyn_;
+ return this->rela_dyn_;
}
// A class to handle the PLT data.
Output_data_plt_x86_64::Output_data_plt_x86_64(Layout* layout,
Output_data_space* got_plt)
- // TODO(csilvers): do we really need an alignment of 8?
: Output_section_data(8), got_plt_(got_plt), count_(0)
{
this->rel_ = new Reloc_section();
Layout* layout,
Sized_relobj<64, false>* object,
unsigned int data_shndx, Symbol* gsym,
- const elfcpp::Rela<64, false>& rel)
+ const elfcpp::Rela<64, false>& rela)
{
Sized_symbol<64>* ssym;
ssym = symtab->get_sized_symbol SELECT_SIZE_NAME(64) (gsym
// symbol, then we will emit the relocation.
if (this->copy_relocs_ == NULL)
this->copy_relocs_ = new Copy_relocs<64, false>();
- this->copy_relocs_->save(ssym, object, data_shndx, rel);
+ this->copy_relocs_->save(ssym, object, data_shndx, rela);
}
else
{
off_t offset = dynbss_size;
dynbss->set_space_size(dynbss_size + symsize);
- // Define the symbol in the .dynbss section.
- symtab->define_in_output_data(this, ssym->name(), ssym->version(),
- dynbss, offset, symsize, ssym->type(),
- ssym->binding(), ssym->visibility(),
- ssym->nonvis(), false, false);
+ symtab->define_with_copy_reloc(this, ssym, dynbss, offset);
// Add the COPY reloc.
- ssym->set_needs_dynsym_entry();
- Reloc_section* rel_dyn = this->rel_dyn_section(layout);
- rel_dyn->add_global(ssym, elfcpp::R_X86_64_COPY, dynbss, offset, 0);
+ Reloc_section* rela_dyn = this->rela_dyn_section(layout);
+ rela_dyn->add_global(ssym, elfcpp::R_X86_64_COPY, dynbss, offset, 0);
}
}
// symbol. IS_FINAL is true if the final address of this symbol is
// known at link time.
-unsigned int
+tls::Tls_optimization
Target_x86_64::optimize_tls_reloc(bool is_final, int r_type)
{
// If we are generating a shared library, then we can't do anything
// in the linker.
if (parameters->output_is_shared())
- return r_type;
+ return tls::TLSOPT_NONE;
switch (r_type)
{
case elfcpp::R_X86_64_TLSGD:
- case elfcpp::R_X86_64_GOTPC32_TLSDESC: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSDESC_CALL: // TODO(csilvers): correct?
- // These are Global-Dynamic which permits fully general TLS
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC:
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ // These are General-Dynamic which permits fully general TLS
// access. Since we know that we are generating an executable,
// we can convert this to Initial-Exec. If we also know that
// this is a local symbol, we can further switch to Local-Exec.
if (is_final)
- return elfcpp::R_X86_64_TPOFF32;
- return elfcpp::R_X86_64_GOTTPOFF; // used for Initial-exec
+ return tls::TLSOPT_TO_LE;
+ return tls::TLSOPT_TO_IE;
case elfcpp::R_X86_64_TLSLD:
// This is Local-Dynamic, which refers to a local symbol in the
// dynamic TLS block. Since we know that we generating an
// executable, we can switch to Local-Exec.
- return elfcpp::R_X86_64_TPOFF32;
+ return tls::TLSOPT_TO_LE;
+
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ // Another Local-Dynamic reloc.
+ return tls::TLSOPT_TO_LE;
case elfcpp::R_X86_64_GOTTPOFF:
// These are Initial-Exec relocs which get the thread offset
// local symbol, we can switch to Local-Exec, which links the
// thread offset into the instruction.
if (is_final)
- return elfcpp::R_X86_64_TPOFF32;
- return r_type;
+ return tls::TLSOPT_TO_LE;
+ return tls::TLSOPT_NONE;
case elfcpp::R_X86_64_TPOFF32:
// When we already have Local-Exec, there is nothing further we
// can do.
- return r_type;
+ return tls::TLSOPT_NONE;
default:
gold_unreachable();
}
}
+// Report an unsupported relocation against a local symbol.
+
+void
+Target_x86_64::Scan::unsupported_reloc_local(Sized_relobj<64, false>* object,
+ unsigned int r_type)
+{
+ gold_error(_("%s: unsupported reloc %u against local symbol"),
+ object->name().c_str(), r_type);
+}
+
// Scan a relocation for a local symbol.
inline void
Layout* layout,
Target_x86_64* target,
Sized_relobj<64, false>* object,
- unsigned int,
- const elfcpp::Rela<64, false>&,
+ unsigned int data_shndx,
+ const elfcpp::Rela<64, false>& reloc,
unsigned int r_type,
const elfcpp::Sym<64, false>&)
{
break;
case elfcpp::R_X86_64_64:
+ // If building a shared library (or a position-independent
+ // executable), we need to create a dynamic relocation for
+ // this location. The relocation applied at link time will
+ // apply the link-time value, so we flag the location with
+ // an R_386_RELATIVE relocation so the dynamic loader can
+ // relocate it easily.
+ if (parameters->output_is_position_independent())
+ {
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ rela_dyn->add_local(object, 0, elfcpp::R_X86_64_RELATIVE,
+ data_shndx, reloc.get_r_offset(), 0);
+ }
+ break;
+
case elfcpp::R_X86_64_32:
case elfcpp::R_X86_64_32S:
case elfcpp::R_X86_64_16:
case elfcpp::R_X86_64_8:
- // FIXME: If we are generating a shared object we need to copy
- // this relocation into the object.
- gold_assert(!parameters->output_is_shared());
+ // If building a shared library (or a position-independent
+ // executable), we need to create a dynamic relocation for
+ // this location. The relocation applied at link time will
+ // apply the link-time value, so we flag the location with
+ // an R_386_RELATIVE relocation so the dynamic loader can
+ // relocate it easily.
+ if (parameters->output_is_position_independent())
+ {
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ unsigned int r_sym = elfcpp::elf_r_sym<64>(reloc.get_r_info());
+ rela_dyn->add_local(object, r_sym, r_type, data_shndx,
+ reloc.get_r_offset(),
+ reloc.get_r_addend());
+ }
break;
case elfcpp::R_X86_64_PC64:
case elfcpp::R_X86_64_PC8:
break;
- case elfcpp::R_X86_64_GOTPC32: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_PLT32:
+ // Since we know this is a local symbol, we can handle this as a
+ // PC32 reloc.
+ break;
+
+ case elfcpp::R_X86_64_GOTPC32:
case elfcpp::R_X86_64_GOTOFF64:
- case elfcpp::R_X86_64_GOTPC64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_PLTOFF64: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_GOTPC64:
+ case elfcpp::R_X86_64_PLTOFF64:
// We need a GOT section.
target->got_section(symtab, layout);
+ // For PLTOFF64, we'd normally want a PLT section, but since we
+ // know this is a local symbol, no PLT is needed.
+ break;
+
+ case elfcpp::R_X86_64_GOT64:
+ case elfcpp::R_X86_64_GOT32:
+ case elfcpp::R_X86_64_GOTPCREL64:
+ case elfcpp::R_X86_64_GOTPCREL:
+ case elfcpp::R_X86_64_GOTPLT64:
+ {
+ // The symbol requires a GOT entry.
+ Output_data_got<64, false>* got = target->got_section(symtab, layout);
+ unsigned int r_sym = elfcpp::elf_r_sym<64>(reloc.get_r_info());
+ if (got->add_local(object, r_sym))
+ {
+ // If we are generating a shared object, we need to add a
+ // dynamic RELATIVE relocation for this symbol.
+ if (parameters->output_is_position_independent())
+ {
+ // FIXME: R_X86_64_RELATIVE assumes a 64-bit relocation.
+ gold_assert(r_type != elfcpp::R_X86_64_GOT32);
+
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ rela_dyn->add_local(object, 0, elfcpp::R_X86_64_RELATIVE,
+ data_shndx, reloc.get_r_offset(), 0);
+ }
+ }
+ // For GOTPLT64, we'd normally want a PLT section, but since
+ // we know this is a local symbol, no PLT is needed.
+ }
break;
case elfcpp::R_X86_64_COPY:
// These are outstanding tls relocs, which are unexpected when linking
case elfcpp::R_X86_64_TPOFF64:
case elfcpp::R_X86_64_DTPMOD64:
- case elfcpp::R_X86_64_DTPOFF64:
- case elfcpp::R_X86_64_DTPOFF32:
case elfcpp::R_X86_64_TLSDESC:
- fprintf(stderr, _("%s: %s: unexpected reloc %u in object file\n"),
- program_name, object->name().c_str(), r_type);
- gold_exit(false);
+ gold_error(_("%s: unexpected reloc %u in object file"),
+ object->name().c_str(), r_type);
break;
// These are initial tls relocs, which are expected when linking
- case elfcpp::R_X86_64_TLSGD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSLD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTTPOFF: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TPOFF32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTPC32_TLSDESC: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSDESC_CALL: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_TLSGD: // Global-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC: // Global-dynamic (from ~oliva url)
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ case elfcpp::R_X86_64_TPOFF32: // Local-exec
{
bool output_is_shared = parameters->output_is_shared();
- r_type = Target_x86_64::optimize_tls_reloc(!output_is_shared, r_type);
+ const tls::Tls_optimization optimized_type
+ = Target_x86_64::optimize_tls_reloc(!output_is_shared, r_type);
switch (r_type)
{
+ case elfcpp::R_X86_64_TLSGD: // General-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC:
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ // FIXME: If not relaxing to LE, we need to generate
+ // DTPMOD64 and DTPOFF64 relocs.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_local(object, r_type);
+ break;
+
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ // FIXME: If not relaxing to LE, we need to generate a
+ // DTPMOD64 reloc.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_local(object, r_type);
+ break;
+
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ // FIXME: If not relaxing to LE, we need to generate a
+ // TPOFF64 reloc.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_local(object, r_type);
+ break;
+
case elfcpp::R_X86_64_TPOFF32: // Local-exec
// FIXME: If generating a shared object, we need to copy
// this relocation into the object.
gold_assert(!output_is_shared);
break;
- case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
- case elfcpp::R_X86_64_TLSGD: // General Dynamic
- case elfcpp::R_X86_64_TLSLD: // Local Dynamic
- case elfcpp::R_X86_64_GOTPC32_TLSDESC:
- case elfcpp::R_X86_64_TLSDESC_CALL:
- fprintf(stderr,
- _("%s: %s: unsupported reloc %u against local symbol\n"),
- program_name, object->name().c_str(), r_type);
- break;
+ default:
+ gold_unreachable();
}
}
break;
- case elfcpp::R_X86_64_GOT64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOT32:
- case elfcpp::R_X86_64_GOTPCREL64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTPCREL:
- case elfcpp::R_X86_64_GOTPLT64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_PLT32:
- case elfcpp::R_X86_64_SIZE32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_SIZE64: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_SIZE32:
+ case elfcpp::R_X86_64_SIZE64:
default:
- fprintf(stderr, _("%s: %s: unsupported reloc %u against local symbol\n"),
- program_name, object->name().c_str(), r_type);
+ gold_error(_("%s: unsupported reloc %u against local symbol"),
+ object->name().c_str(), r_type);
break;
}
}
+// Report an unsupported relocation against a global symbol.
+
+void
+Target_x86_64::Scan::unsupported_reloc_global(Sized_relobj<64, false>* object,
+ unsigned int r_type,
+ Symbol* gsym)
+{
+ gold_error(_("%s: unsupported reloc %u against global symbol %s"),
+ object->name().c_str(), r_type, gsym->name());
+}
+
// Scan a relocation for a global symbol.
inline void
break;
case elfcpp::R_X86_64_64:
- case elfcpp::R_X86_64_PC64:
case elfcpp::R_X86_64_32:
case elfcpp::R_X86_64_32S:
- case elfcpp::R_X86_64_PC32:
case elfcpp::R_X86_64_16:
- case elfcpp::R_X86_64_PC16:
case elfcpp::R_X86_64_8:
- case elfcpp::R_X86_64_PC8:
- // FIXME: If we are generating a shared object we may need to
- // copy this relocation into the object. If this symbol is
- // defined in a shared object, we may need to copy this
- // relocation in order to avoid a COPY relocation.
- gold_assert(!parameters->output_is_shared());
-
- if (gsym->is_from_dynobj())
- {
- // This symbol is defined in a dynamic object. If it is a
- // function, we make a PLT entry. Otherwise we need to
- // either generate a COPY reloc or copy this reloc.
- if (gsym->type() == elfcpp::STT_FUNC)
- {
- target->make_plt_entry(symtab, layout, gsym);
-
- // If this is not a PC relative reference, then we may
- // be taking the address of the function. In that case
- // we need to set the entry in the dynamic symbol table
- // to the address of the PLT entry.
- if (r_type != elfcpp::R_X86_64_PC64
- && r_type != elfcpp::R_X86_64_PC32
- && r_type != elfcpp::R_X86_64_PC16
- && r_type != elfcpp::R_X86_64_PC8)
- gsym->set_needs_dynsym_value();
- }
- else
- target->copy_reloc(&options, symtab, layout, object, data_shndx,
- gsym, reloc);
- }
+ {
+ // Make a PLT entry if necessary.
+ if (gsym->needs_plt_entry())
+ {
+ target->make_plt_entry(symtab, layout, gsym);
+ // Since this is not a PC-relative relocation, we may be
+ // taking the address of a function. In that case we need to
+ // set the entry in the dynamic symbol table to the address of
+ // the PLT entry.
+ if (gsym->is_from_dynobj())
+ gsym->set_needs_dynsym_value();
+ }
+ // Make a dynamic relocation if necessary.
+ if (gsym->needs_dynamic_reloc(true, false))
+ {
+ if (target->may_need_copy_reloc(gsym))
+ {
+ target->copy_reloc(&options, symtab, layout, object, data_shndx,
+ gsym, reloc);
+ }
+ else if (r_type == elfcpp::R_X86_64_64
+ && gsym->can_use_relative_reloc(false))
+ {
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ rela_dyn->add_local(object, 0, elfcpp::R_X86_64_RELATIVE,
+ data_shndx,
+ reloc.get_r_offset(), 0);
+ }
+ else
+ {
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ rela_dyn->add_global(gsym, r_type, object, data_shndx,
+ reloc.get_r_offset(),
+ reloc.get_r_addend());
+ }
+ }
+ }
+ break;
+ case elfcpp::R_X86_64_PC64:
+ case elfcpp::R_X86_64_PC32:
+ case elfcpp::R_X86_64_PC16:
+ case elfcpp::R_X86_64_PC8:
+ {
+ // Make a PLT entry if necessary.
+ if (gsym->needs_plt_entry())
+ target->make_plt_entry(symtab, layout, gsym);
+ // Make a dynamic relocation if necessary.
+ bool is_function_call = (gsym->type() == elfcpp::STT_FUNC);
+ if (gsym->needs_dynamic_reloc(true, is_function_call))
+ {
+ if (target->may_need_copy_reloc(gsym))
+ {
+ target->copy_reloc(&options, symtab, layout, object, data_shndx,
+ gsym, reloc);
+ }
+ else
+ {
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ rela_dyn->add_global(gsym, r_type, object, data_shndx,
+ reloc.get_r_offset(),
+ reloc.get_r_addend());
+ }
+ }
+ }
break;
case elfcpp::R_X86_64_GOT64:
// dynamic relocation for it.
if (!gsym->final_value_is_known())
{
- Reloc_section* rel_dyn = target->rel_dyn_section(layout);
- rel_dyn->add_global(gsym, elfcpp::R_X86_64_GLOB_DAT, got,
- gsym->got_offset(), 0);
+ Reloc_section* rela_dyn = target->rela_dyn_section(layout);
+ if (gsym->is_from_dynobj()
+ || gsym->is_preemptible())
+ rela_dyn->add_global(gsym, elfcpp::R_X86_64_GLOB_DAT, got,
+ gsym->got_offset(), 0);
+ else
+ {
+ rela_dyn->add_local(object, 0, elfcpp::R_X86_64_RELATIVE,
+ got, gsym->got_offset(), 0);
+ // Make sure we write the link-time value to the GOT.
+ gsym->set_needs_value_in_got();
+ }
}
}
+ // For GOTPLT64, we also need a PLT entry (but only if the
+ // symbol is not fully resolved).
+ if (r_type == elfcpp::R_X86_64_GOTPLT64
+ && !gsym->final_value_is_known())
+ target->make_plt_entry(symtab, layout, gsym);
}
break;
// Otherwise we need a PLT entry.
if (gsym->final_value_is_known())
break;
+ // If building a shared library, we can also skip the PLT entry
+ // if the symbol is defined in the output file and is protected
+ // or hidden.
+ if (gsym->is_defined()
+ && !gsym->is_from_dynobj()
+ && !gsym->is_preemptible())
+ break;
target->make_plt_entry(symtab, layout, gsym);
break;
- case elfcpp::R_X86_64_GOTPC32: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_GOTPC32:
case elfcpp::R_X86_64_GOTOFF64:
- case elfcpp::R_X86_64_GOTPC64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_PLTOFF64: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_GOTPC64:
+ case elfcpp::R_X86_64_PLTOFF64:
// We need a GOT section.
target->got_section(symtab, layout);
+ // For PLTOFF64, we also need a PLT entry (but only if the
+ // symbol is not fully resolved).
+ if (r_type == elfcpp::R_X86_64_PLTOFF64
+ && !gsym->final_value_is_known())
+ target->make_plt_entry(symtab, layout, gsym);
break;
case elfcpp::R_X86_64_COPY:
// These are outstanding tls relocs, which are unexpected when linking
case elfcpp::R_X86_64_TPOFF64:
case elfcpp::R_X86_64_DTPMOD64:
- case elfcpp::R_X86_64_DTPOFF64:
- case elfcpp::R_X86_64_DTPOFF32:
case elfcpp::R_X86_64_TLSDESC:
- fprintf(stderr, _("%s: %s: unexpected reloc %u in object file\n"),
- program_name, object->name().c_str(), r_type);
- gold_exit(false);
+ gold_error(_("%s: unexpected reloc %u in object file"),
+ object->name().c_str(), r_type);
break;
// These are initial tls relocs, which are expected for global()
- case elfcpp::R_X86_64_TLSGD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSLD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTTPOFF: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TPOFF32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTPC32_TLSDESC: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSDESC_CALL: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_TLSGD: // Global-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC: // Global-dynamic (from ~oliva url)
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ case elfcpp::R_X86_64_TPOFF32: // Local-exec
{
const bool is_final = gsym->final_value_is_known();
- r_type = Target_x86_64::optimize_tls_reloc(is_final, r_type);
+ const tls::Tls_optimization optimized_type
+ = Target_x86_64::optimize_tls_reloc(is_final, r_type);
switch (r_type)
{
+ case elfcpp::R_X86_64_TLSGD: // General-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC:
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ // FIXME: If not relaxing to LE, we need to generate
+ // DTPMOD64 and DTPOFF64, or TLSDESC, relocs.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_global(object, r_type, gsym);
+ break;
+
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ // FIXME: If not relaxing to LE, we need to generate a
+ // DTPMOD64 reloc.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_global(object, r_type, gsym);
+ break;
+
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ // FIXME: If not relaxing to LE, we need to generate a
+ // TPOFF64 reloc.
+ if (optimized_type != tls::TLSOPT_TO_LE)
+ unsupported_reloc_global(object, r_type, gsym);
+ break;
+
case elfcpp::R_X86_64_TPOFF32: // Local-exec
// FIXME: If generating a shared object, we need to copy
// this relocation into the object.
- gold_assert(!parameters->output_is_shared());
+ gold_assert(is_final);
break;
- case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
- case elfcpp::R_X86_64_TLSGD: // General Dynamic
- case elfcpp::R_X86_64_TLSLD: // Local Dynamic
- case elfcpp::R_X86_64_GOTPC32_TLSDESC:
- case elfcpp::R_X86_64_TLSDESC_CALL:
- fprintf(stderr,
- _("%s: %s: unsupported reloc %u "
- "against global symbol %s\n"),
- program_name, object->name().c_str(), r_type,
- gsym->name());
- break;
+ default:
+ gold_unreachable();
}
}
break;
- case elfcpp::R_X86_64_SIZE32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_SIZE64: // TODO(csilvers): correct?
+ case elfcpp::R_X86_64_SIZE32:
+ case elfcpp::R_X86_64_SIZE64:
default:
- fprintf(stderr,
- _("%s: %s: unsupported reloc %u against global symbol %s\n"),
- program_name, object->name().c_str(), r_type, gsym->name());
+ gold_error(_("%s: unsupported reloc %u against global symbol %s"),
+ object->name().c_str(), r_type, gsym->name());
break;
}
}
unsigned int sh_type,
const unsigned char* prelocs,
size_t reloc_count,
+ Output_section* output_section,
+ bool needs_special_offset_handling,
size_t local_symbol_count,
- const unsigned char* plocal_symbols,
- Symbol** global_symbols)
+ const unsigned char* plocal_symbols)
{
if (sh_type == elfcpp::SHT_REL)
{
- fprintf(stderr, _("%s: %s: unsupported REL reloc section\n"),
- program_name, object->name().c_str());
- gold_exit(false);
+ gold_error(_("%s: unsupported REL reloc section"),
+ object->name().c_str());
+ return;
}
gold::scan_relocs<64, false, Target_x86_64, elfcpp::SHT_RELA,
data_shndx,
prelocs,
reloc_count,
+ output_section,
+ needs_special_offset_handling,
local_symbol_count,
- plocal_symbols,
- global_symbols);
+ plocal_symbols);
}
// Finalize the sections.
odyn->add_constant(elfcpp::DT_PLTREL, elfcpp::DT_RELA);
}
- if (this->rel_dyn_ != NULL)
+ if (this->rela_dyn_ != NULL)
{
- const Output_data* od = this->rel_dyn_;
+ const Output_data* od = this->rela_dyn_;
odyn->add_section_address(elfcpp::DT_RELA, od);
odyn->add_section_size(elfcpp::DT_RELASZ, od);
odyn->add_constant(elfcpp::DT_RELAENT,
return;
if (this->copy_relocs_->any_to_emit())
{
- Reloc_section* rel_dyn = this->rel_dyn_section(layout);
- this->copy_relocs_->emit(rel_dyn);
+ Reloc_section* rela_dyn = this->rela_dyn_section(layout);
+ this->copy_relocs_->emit(rela_dyn);
}
delete this->copy_relocs_;
this->copy_relocs_ = NULL;
Target_x86_64::Relocate::relocate(const Relocate_info<64, false>* relinfo,
Target_x86_64* target,
size_t relnum,
- const elfcpp::Rela<64, false>& rel,
+ const elfcpp::Rela<64, false>& rela,
unsigned int r_type,
const Sized_symbol<64>* gsym,
const Symbol_value<64>* psymval,
{
if (r_type != elfcpp::R_X86_64_PLT32
|| gsym == NULL
- || strcmp(gsym->name(), "___tls_get_addr") != 0)
+ || strcmp(gsym->name(), "__tls_get_addr") != 0)
{
- fprintf(stderr, _("%s: %s: missing expected TLS relocation\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str());
- gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("missing expected TLS relocation"));
+ }
+ else
+ {
+ this->skip_call_tls_get_addr_ = false;
+ return false;
}
-
- this->skip_call_tls_get_addr_ = false;
-
- return false;
}
// Pick the value to use for symbols defined in shared objects.
Symbol_value<64> symval;
- if (gsym != NULL && gsym->is_from_dynobj() && gsym->has_plt_offset())
+ if (gsym != NULL
+ && (gsym->is_from_dynobj()
+ || (parameters->output_is_shared()
+ && gsym->is_preemptible()))
+ && gsym->has_plt_offset())
{
symval.set_output_value(target->plt_section()->address()
+ gsym->plt_offset());
}
const Sized_relobj<64, false>* object = relinfo->object;
- const elfcpp::Elf_Xword addend = rel.get_r_addend();
+ const elfcpp::Elf_Xword addend = rela.get_r_addend();
+
+ // Get the GOT offset if needed.
+ // The GOT pointer points to the end of the GOT section.
+ // We need to subtract the size of the GOT section to get
+ // the actual offset to use in the relocation.
+ bool have_got_offset = false;
+ unsigned int got_offset = 0;
+ switch (r_type)
+ {
+ case elfcpp::R_X86_64_GOT32:
+ case elfcpp::R_X86_64_GOT64:
+ case elfcpp::R_X86_64_GOTPLT64:
+ case elfcpp::R_X86_64_GOTPCREL:
+ case elfcpp::R_X86_64_GOTPCREL64:
+ if (gsym != NULL)
+ {
+ gold_assert(gsym->has_got_offset());
+ got_offset = gsym->got_offset() - target->got_size();
+ }
+ else
+ {
+ unsigned int r_sym = elfcpp::elf_r_sym<64>(rela.get_r_info());
+ got_offset = object->local_got_offset(r_sym) - target->got_size();
+ }
+ have_got_offset = true;
+ break;
+
+ default:
+ break;
+ }
switch (r_type)
{
break;
case elfcpp::R_X86_64_PLT32:
- gold_assert(gsym->has_plt_offset()
+ gold_assert(gsym == NULL
+ || gsym->has_plt_offset()
|| gsym->final_value_is_known());
+ // Note: while this code looks the same as for R_X86_64_PC32, it
+ // behaves differently because psymval was set to point to
+ // the PLT entry, rather than the symbol, in Scan::global().
Relocate_functions<64, false>::pcrela32(view, object, psymval, addend,
address);
break;
+ case elfcpp::R_X86_64_PLTOFF64:
+ {
+ gold_assert(gsym);
+ gold_assert(gsym->has_plt_offset()
+ || gsym->final_value_is_known());
+ elfcpp::Elf_types<64>::Elf_Addr got_address;
+ got_address = target->got_section(NULL, NULL)->address();
+ Relocate_functions<64, false>::rela64(view, object, psymval,
+ addend - got_address);
+ }
+
case elfcpp::R_X86_64_GOT32:
- // Local GOT offsets not yet supported.
- gold_assert(gsym);
- gold_assert(gsym->has_got_offset());
- Relocate_functions<64, false>::rela32(view, gsym->got_offset(), addend);
+ gold_assert(have_got_offset);
+ Relocate_functions<64, false>::rela32(view, got_offset, addend);
break;
case elfcpp::R_X86_64_GOTPC32:
{
gold_assert(gsym);
elfcpp::Elf_types<64>::Elf_Addr value;
- value = target->got_section(NULL, NULL)->address();
+ value = target->got_plt_section()->address();
Relocate_functions<64, false>::pcrela32(view, value, addend, address);
}
break;
case elfcpp::R_X86_64_GOT64:
// The ABI doc says "Like GOT64, but indicates a PLT entry is needed."
// Since we always add a PLT entry, this is equivalent.
- case elfcpp::R_X86_64_GOTPLT64: // TODO(csilvers): correct?
- // Local GOT offsets not yet supported.
- gold_assert(gsym);
- gold_assert(gsym->has_got_offset());
- Relocate_functions<64, false>::rela64(view, gsym->got_offset(), addend);
+ case elfcpp::R_X86_64_GOTPLT64:
+ gold_assert(have_got_offset);
+ Relocate_functions<64, false>::rela64(view, got_offset, addend);
break;
case elfcpp::R_X86_64_GOTPC64:
{
gold_assert(gsym);
elfcpp::Elf_types<64>::Elf_Addr value;
- value = target->got_section(NULL, NULL)->address();
+ value = target->got_plt_section()->address();
Relocate_functions<64, false>::pcrela64(view, value, addend, address);
}
break;
{
elfcpp::Elf_types<64>::Elf_Addr value;
value = (psymval->value(object, 0)
- - target->got_section(NULL, NULL)->address());
+ - target->got_plt_section()->address());
Relocate_functions<64, false>::rela64(view, value, addend);
}
break;
case elfcpp::R_X86_64_GOTPCREL:
{
- // Local GOT offsets not yet supported.
- gold_assert(gsym);
- gold_assert(gsym->has_got_offset());
- elfcpp::Elf_types<64>::Elf_Addr value;
- value = (target->got_section(NULL, NULL)->address()
- + gsym->got_offset());
- Relocate_functions<64, false>::pcrela32(view, value, addend, address);
+ gold_assert(have_got_offset);
+ elfcpp::Elf_types<64>::Elf_Addr value;
+ value = target->got_plt_section()->address() + got_offset;
+ Relocate_functions<64, false>::pcrela32(view, value, addend, address);
}
break;
case elfcpp::R_X86_64_GOTPCREL64:
{
- // Local GOT offsets not yet supported.
- gold_assert(gsym);
- gold_assert(gsym->has_got_offset());
- elfcpp::Elf_types<64>::Elf_Addr value;
- value = (target->got_section(NULL, NULL)->address()
- + gsym->got_offset());
- Relocate_functions<64, false>::pcrela64(view, value, addend, address);
+ gold_assert(have_got_offset);
+ elfcpp::Elf_types<64>::Elf_Addr value;
+ value = target->got_plt_section()->address() + got_offset;
+ Relocate_functions<64, false>::pcrela64(view, value, addend, address);
}
break;
// These are outstanding tls relocs, which are unexpected when linking
case elfcpp::R_X86_64_TPOFF64:
case elfcpp::R_X86_64_DTPMOD64:
- case elfcpp::R_X86_64_DTPOFF64:
- case elfcpp::R_X86_64_DTPOFF32:
case elfcpp::R_X86_64_TLSDESC:
- fprintf(stderr, _("%s: %s: unexpected reloc %u in object file\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str(),
- r_type);
- gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("unexpected reloc %u in object file"),
+ r_type);
break;
// These are initial tls relocs, which are expected when linking
- case elfcpp::R_X86_64_TLSGD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSLD: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTTPOFF: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TPOFF32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_GOTPC32_TLSDESC: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_TLSDESC_CALL: // TODO(csilvers): correct?
- this->relocate_tls(relinfo, relnum, rel, r_type, gsym, psymval, view,
+ case elfcpp::R_X86_64_TLSGD: // Global-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC: // Global-dynamic (from ~oliva url)
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ case elfcpp::R_X86_64_DTPOFF32:
+ case elfcpp::R_X86_64_DTPOFF64:
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ case elfcpp::R_X86_64_TPOFF32: // Local-exec
+ this->relocate_tls(relinfo, relnum, rela, r_type, gsym, psymval, view,
address, view_size);
break;
- case elfcpp::R_X86_64_SIZE32: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_SIZE64: // TODO(csilvers): correct?
- case elfcpp::R_X86_64_PLTOFF64: // TODO(csilvers): implement me!
+ case elfcpp::R_X86_64_SIZE32:
+ case elfcpp::R_X86_64_SIZE64:
default:
- fprintf(stderr, _("%s: %s: unsupported reloc %u\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str(),
- r_type);
- // gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("unsupported reloc %u"),
+ r_type);
break;
}
inline void
Target_x86_64::Relocate::relocate_tls(const Relocate_info<64, false>* relinfo,
size_t relnum,
- const elfcpp::Rela<64, false>& rel,
+ const elfcpp::Rela<64, false>& rela,
unsigned int r_type,
const Sized_symbol<64>* gsym,
const Symbol_value<64>* psymval,
Output_segment* tls_segment = relinfo->layout->tls_segment();
if (tls_segment == NULL)
{
- fprintf(stderr, _("%s: %s: TLS reloc but no TLS segment\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str());
- gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("TLS reloc but no TLS segment"));
+ return;
}
elfcpp::Elf_types<64>::Elf_Addr value = psymval->value(relinfo->object, 0);
const bool is_final = (gsym == NULL
- ? !parameters->output_is_shared()
+ ? !parameters->output_is_position_independent()
: gsym->final_value_is_known());
- const unsigned int opt_r_type =
- Target_x86_64::optimize_tls_reloc(is_final, r_type);
+ const tls::Tls_optimization optimized_type
+ = Target_x86_64::optimize_tls_reloc(is_final, r_type);
switch (r_type)
{
- case elfcpp::R_X86_64_TPOFF32: // Local-exec reloc
- value = value - (tls_segment->vaddr() + tls_segment->memsz());
- Relocate_functions<64, false>::rel32(view, value);
- break;
-
- case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec reloc
- if (opt_r_type == elfcpp::R_X86_64_TPOFF32)
+ case elfcpp::R_X86_64_TLSGD: // Global-dynamic
+ case elfcpp::R_X86_64_GOTPC32_TLSDESC: // Global-dynamic (from ~oliva url)
+ case elfcpp::R_X86_64_TLSDESC_CALL:
+ if (optimized_type == tls::TLSOPT_TO_LE)
{
- Target_x86_64::Relocate::tls_ie_to_le(relinfo, relnum, tls_segment,
- rel, r_type, value, view,
- view_size);
+ this->tls_gd_to_le(relinfo, relnum, tls_segment,
+ rela, r_type, value, view,
+ view_size);
break;
}
- fprintf(stderr, _("%s: %s: unsupported reloc type %u\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str(),
- r_type);
- // gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("unsupported reloc %u"), r_type);
break;
- case elfcpp::R_X86_64_TLSGD:
- if (opt_r_type == elfcpp::R_X86_64_TPOFF32)
+ case elfcpp::R_X86_64_TLSLD: // Local-dynamic
+ if (optimized_type == tls::TLSOPT_TO_LE)
+ {
+ this->tls_ld_to_le(relinfo, relnum, tls_segment, rela, r_type,
+ value, view, view_size);
+ break;
+ }
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("unsupported reloc %u"), r_type);
+ break;
+
+ case elfcpp::R_X86_64_DTPOFF32:
+ if (optimized_type == tls::TLSOPT_TO_LE)
+ value = value - (tls_segment->vaddr() + tls_segment->memsz());
+ else
+ value = value - tls_segment->vaddr();
+ Relocate_functions<64, false>::rel32(view, value);
+ break;
+
+ case elfcpp::R_X86_64_DTPOFF64:
+ if (optimized_type == tls::TLSOPT_TO_LE)
+ value = value - (tls_segment->vaddr() + tls_segment->memsz());
+ else
+ value = value - tls_segment->vaddr();
+ Relocate_functions<64, false>::rel64(view, value);
+ break;
+
+ case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec
+ if (optimized_type == tls::TLSOPT_TO_LE)
{
- this->tls_gd_to_le(relinfo, relnum, tls_segment,
- rel, r_type, value, view,
- view_size);
+ Target_x86_64::Relocate::tls_ie_to_le(relinfo, relnum, tls_segment,
+ rela, r_type, value, view,
+ view_size);
break;
}
- fprintf(stderr, _("%s: %s: unsupported reloc %u\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str(),
- r_type);
- // gold_exit(false);
+ gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
+ _("unsupported reloc type %u"),
+ r_type);
break;
- case elfcpp::R_X86_64_TLSLD:
- fprintf(stderr, _("%s: %s: unsupported reloc %u\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str(),
- r_type);
- // gold_exit(false);
+ case elfcpp::R_X86_64_TPOFF32: // Local-exec
+ value = value - (tls_segment->vaddr() + tls_segment->memsz());
+ Relocate_functions<64, false>::rel32(view, value);
break;
}
}
-// Do a relocation in which we convert a TLS Initial-Exec to a
+// Do a relocation in which we convert a TLS General-Dynamic to a
// Local-Exec.
-// TODO(csilvers): verify this is right.
inline void
-Target_x86_64::Relocate::tls_ie_to_le(const Relocate_info<64, false>* relinfo,
+Target_x86_64::Relocate::tls_gd_to_le(const Relocate_info<64, false>* relinfo,
size_t relnum,
Output_segment* tls_segment,
- const elfcpp::Rela<64, false>& rel,
+ const elfcpp::Rela<64, false>& rela,
unsigned int,
elfcpp::Elf_types<64>::Elf_Addr value,
unsigned char* view,
off_t view_size)
{
- // We have to actually change the instructions, which means that we
- // need to examine the opcodes to figure out which instruction we
- // are looking at.
+ // .byte 0x66; leaq foo@tlsgd(%rip),%rdi;
+ // .word 0x6666; rex64; call __tls_get_addr
+ // ==> movq %fs:0,%rax; leaq x@tpoff(%rax),%rax
- // movl %gs:XX,%eax ==> movl $YY,%eax
- // movl %gs:XX,%reg ==> movl $YY,%reg
- // addl %gs:XX,%reg ==> addl $YY,%reg
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel, view_size, -1);
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel, view_size, 4);
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, -4);
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 12);
- unsigned char op1 = view[-1];
- if (op1 == 0xa1)
- {
- // movl XX,%eax ==> movl $YY,%eax
- view[-1] = 0xb8;
- }
- else
- {
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel,
- view_size, -2);
+ tls::check_tls(relinfo, relnum, rela.get_r_offset(),
+ (memcmp(view - 4, "\x66\x48\x8d\x3d", 4) == 0));
+ tls::check_tls(relinfo, relnum, rela.get_r_offset(),
+ (memcmp(view + 4, "\x66\x66\x48\xe8", 4) == 0));
- unsigned char op2 = view[-2];
- if (op2 == 0x8b)
- {
- // movl XX,%reg ==> movl $YY,%reg
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- (op1 & 0xc7) == 0x05);
- view[-2] = 0xc7;
- view[-1] = 0xc0 | ((op1 >> 3) & 7);
- }
- else if (op2 == 0x03)
- {
- // addl XX,%reg ==> addl $YY,%reg
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- (op1 & 0xc7) == 0x05);
- view[-2] = 0x81;
- view[-1] = 0xc0 | ((op1 >> 3) & 7);
- }
- else
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel, 0);
- }
+ memcpy(view - 4, "\x64\x48\x8b\x04\x25\0\0\0\0\x48\x8d\x80\0\0\0\0", 16);
value = value - (tls_segment->vaddr() + tls_segment->memsz());
- Relocate_functions<64, false>::rel32(view, value);
-}
+ Relocate_functions<64, false>::rela32(view + 8, value, 0);
-// Do a relocation in which we convert a TLS Global-Dynamic to a
-// Local-Exec.
-// TODO(csilvers): verify this is right.
+ // The next reloc should be a PLT32 reloc against __tls_get_addr.
+ // We can skip it.
+ this->skip_call_tls_get_addr_ = true;
+}
inline void
-Target_x86_64::Relocate::tls_gd_to_le(const Relocate_info<64, false>* relinfo,
+Target_x86_64::Relocate::tls_ld_to_le(const Relocate_info<64, false>* relinfo,
size_t relnum,
- Output_segment* tls_segment,
- const elfcpp::Rela<64, false>& rel,
+ Output_segment*,
+ const elfcpp::Rela<64, false>& rela,
unsigned int,
- elfcpp::Elf_types<64>::Elf_Addr value,
+ elfcpp::Elf_types<64>::Elf_Addr,
unsigned char* view,
off_t view_size)
{
- // leal foo(,%reg,1),%eax; call ___tls_get_addr
- // ==> movl %gs,0,%eax; subl $foo@tpoff,%eax
- // leal foo(%reg),%eax; call ___tls_get_addr
- // ==> movl %gs:0,%eax; subl $foo@tpoff,%eax
-
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel, view_size, -2);
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel, view_size, 9);
-
- unsigned char op1 = view[-1];
- unsigned char op2 = view[-2];
+ // leaq foo@tlsld(%rip),%rdi; call __tls_get_addr@plt;
+ // ... leq foo@dtpoff(%rax),%reg
+ // ==> .word 0x6666; .byte 0x66; movq %fs:0,%rax ... leaq x@tpoff(%rax),%rdx
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- op2 == 0x8d || op2 == 0x04);
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- view[4] == 0xe8);
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, -3);
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 9);
- int roff = 5;
+ tls::check_tls(relinfo, relnum, rela.get_r_offset(),
+ view[-3] == 0x48 && view[-2] == 0x8d && view[-1] == 0x3d);
- if (op2 == 0x04)
- {
- Target_x86_64::Relocate::check_range(relinfo, relnum, rel, view_size, -3);
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- view[-3] == 0x8d);
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- ((op1 & 0xc7) == 0x05
- && op1 != (4 << 3)));
- memcpy(view - 3, "\x65\xa1\0\0\0\0\x81\xe8\0\0\0", 12);
- }
- else
- {
- Target_x86_64::Relocate::check_tls(relinfo, relnum, rel,
- (op1 & 0xf8) == 0x80 && (op1 & 7) != 4);
- if (static_cast<off_t>(rel.get_r_offset() + 9) < view_size
- && view[9] == 0x90)
- {
- // There is a trailing nop. Use the size byte subl.
- memcpy(view - 2, "\x65\xa1\0\0\0\0\x81\xe8\0\0\0", 12);
- roff = 6;
- }
- else
- {
- // Use the five byte subl.
- memcpy(view - 2, "\x65\xa1\0\0\0\0\x2d\0\0\0", 11);
- }
- }
+ tls::check_tls(relinfo, relnum, rela.get_r_offset(), view[4] == 0xe8);
- value = tls_segment->vaddr() + tls_segment->memsz() - value;
- Relocate_functions<64, false>::rel32(view + roff, value);
+ memcpy(view - 3, "\x66\x66\x66\x64\x48\x8b\x04\x25\0\0\0\0", 12);
// The next reloc should be a PLT32 reloc against __tls_get_addr.
// We can skip it.
this->skip_call_tls_get_addr_ = true;
}
-// Check the range for a TLS relocation.
+// Do a relocation in which we convert a TLS Initial-Exec to a
+// Local-Exec.
inline void
-Target_x86_64::Relocate::check_range(const Relocate_info<64, false>* relinfo,
- size_t relnum,
- const elfcpp::Rela<64, false>& rel,
- off_t view_size, off_t off)
+Target_x86_64::Relocate::tls_ie_to_le(const Relocate_info<64, false>* relinfo,
+ size_t relnum,
+ Output_segment* tls_segment,
+ const elfcpp::Rela<64, false>& rela,
+ unsigned int,
+ elfcpp::Elf_types<64>::Elf_Addr value,
+ unsigned char* view,
+ off_t view_size)
{
- off_t offset = rel.get_r_offset() + off;
- if (offset < 0 || offset > view_size)
- {
- fprintf(stderr, _("%s: %s: TLS relocation out of range\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str());
- gold_exit(false);
- }
-}
+ // We need to examine the opcodes to figure out which instruction we
+ // are looking at.
-// Check the validity of a TLS relocation. This is like assert.
+ // movq foo@gottpoff(%rip),%reg ==> movq $YY,%reg
+ // addq foo@gottpoff(%rip),%reg ==> addq $YY,%reg
-inline void
-Target_x86_64::Relocate::check_tls(const Relocate_info<64, false>* relinfo,
- size_t relnum,
- const elfcpp::Rela<64, false>& rel,
- bool valid)
-{
- if (!valid)
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, -3);
+ tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 4);
+
+ unsigned char op1 = view[-3];
+ unsigned char op2 = view[-2];
+ unsigned char op3 = view[-1];
+ unsigned char reg = op3 >> 3;
+
+ if (op2 == 0x8b)
+ {
+ // movq
+ if (op1 == 0x4c)
+ view[-3] = 0x49;
+ view[-2] = 0xc7;
+ view[-1] = 0xc0 | reg;
+ }
+ else if (reg == 4)
+ {
+ // Special handling for %rsp.
+ if (op1 == 0x4c)
+ view[-3] = 0x49;
+ view[-2] = 0x81;
+ view[-1] = 0xc0 | reg;
+ }
+ else
{
- fprintf(stderr,
- _("%s: %s: TLS relocation against invalid instruction\n"),
- program_name,
- relinfo->location(relnum, rel.get_r_offset()).c_str());
- gold_exit(false);
+ // addq
+ if (op1 == 0x4c)
+ view[-3] = 0x4d;
+ view[-2] = 0x8d;
+ view[-1] = 0x80 | reg | (reg << 3);
}
+
+ value = value - (tls_segment->vaddr() + tls_segment->memsz());
+ Relocate_functions<64, false>::rela32(view, value, 0);
}
// Relocate section data.
unsigned int sh_type,
const unsigned char* prelocs,
size_t reloc_count,
+ Output_section* output_section,
+ bool needs_special_offset_handling,
unsigned char* view,
elfcpp::Elf_types<64>::Elf_Addr address,
off_t view_size)
this,
prelocs,
reloc_count,
+ output_section,
+ needs_special_offset_handling,
view,
address,
view_size);