X-Git-Url: http://git.efficios.com/?a=blobdiff_plain;f=gold%2Fx86_64.cc;h=bafd90ef5ef07cea125d08f72330bd2d32a2c544;hb=c1b5c1ebc938b6dc0277363b8c47d75b0b5a621f;hp=255ebb90c376eb6c45b61e98a55c43732d6a32ab;hpb=57b2284c6318841612acef23c60fe7298a1ac91a;p=deliverable%2Fbinutils-gdb.git diff --git a/gold/x86_64.cc b/gold/x86_64.cc index 255ebb90c3..bafd90ef5e 100644 --- a/gold/x86_64.cc +++ b/gold/x86_64.cc @@ -1,6 +1,6 @@ // x86_64.cc -- x86_64 target support for gold. -// Copyright (C) 2006-2014 Free Software Foundation, Inc. +// Copyright (C) 2006-2019 Free Software Foundation, Inc. // Written by Ian Lance Taylor . // This file is part of gold. @@ -179,7 +179,12 @@ class Output_data_plt_x86_64 : public Output_section_data has_irelative_section() const { return this->irelative_rel_ != NULL; } - // Return the number of PLT entries. + // Get count of regular PLT entries. + unsigned int + regular_count() const + { return this->count_; } + + // Return the total number of PLT entries. unsigned int entry_count() const { return this->count_ + this->irelative_count_; } @@ -204,11 +209,13 @@ class Output_data_plt_x86_64 : public Output_section_data // Return the PLT address to use for a global symbol. uint64_t - address_for_global(const Symbol*); + address_for_global(const Symbol* sym) + { return do_address_for_global(sym); } // Return the PLT address to use for a local symbol. uint64_t - address_for_local(const Relobj*, unsigned int symndx); + address_for_local(const Relobj* obj, unsigned int symndx) + { return do_address_for_local(obj, symndx); } // Add .eh_frame information for the PLT. void @@ -216,6 +223,18 @@ class Output_data_plt_x86_64 : public Output_section_data { this->do_add_eh_frame(layout); } protected: + Output_data_got<64, false>* + got() const + { return this->got_; } + + Output_data_got_plt_x86_64* + got_plt() const + { return this->got_plt_; } + + Output_data_space* + got_irelative() const + { return this->got_irelative_; } + // Fill in the first PLT entry. void fill_first_plt_entry(unsigned char* pov, @@ -275,6 +294,14 @@ class Output_data_plt_x86_64 : public Output_section_data unsigned int tlsdesc_got_offset, unsigned int plt_offset) = 0; + // Return the PLT address to use for a global symbol. + virtual uint64_t + do_address_for_global(const Symbol* sym); + + // Return the PLT address to use for a local symbol. + virtual uint64_t + do_address_for_local(const Relobj* obj, unsigned int symndx); + virtual void do_add_eh_frame(Layout* layout) = 0; @@ -403,6 +430,260 @@ class Output_data_plt_x86_64_standard : public Output_data_plt_x86_64 static const unsigned char plt_eh_frame_fde[plt_eh_frame_fde_size]; }; +class Output_data_plt_x86_64_bnd : public Output_data_plt_x86_64<64> +{ + public: + Output_data_plt_x86_64_bnd(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative) + : Output_data_plt_x86_64<64>(layout, plt_entry_size, + got, got_plt, got_irelative), + aplt_offset_(0) + { } + + Output_data_plt_x86_64_bnd(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative, + unsigned int plt_count) + : Output_data_plt_x86_64<64>(layout, plt_entry_size, + got, got_plt, got_irelative, + plt_count), + aplt_offset_(0) + { } + + protected: + virtual unsigned int + do_get_plt_entry_size() const + { return plt_entry_size; } + + // Return the PLT address to use for a global symbol. + uint64_t + do_address_for_global(const Symbol*); + + // Return the PLT address to use for a local symbol. + uint64_t + do_address_for_local(const Relobj*, unsigned int symndx); + + virtual void + do_add_eh_frame(Layout* layout) + { + layout->add_eh_frame_for_plt(this, + this->plt_eh_frame_cie, + this->plt_eh_frame_cie_size, + plt_eh_frame_fde, + plt_eh_frame_fde_size); + } + + virtual void + do_fill_first_plt_entry(unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_addr, + elfcpp::Elf_types<64>::Elf_Addr plt_addr); + + virtual unsigned int + do_fill_plt_entry(unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index); + + virtual void + do_fill_tlsdesc_entry(unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address, + elfcpp::Elf_types<64>::Elf_Addr got_base, + unsigned int tlsdesc_got_offset, + unsigned int plt_offset); + + void + fill_aplt_entry(unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index); + + private: + // Set the final size. + void + set_final_data_size(); + + // Write out the BND PLT data. + void + do_write(Output_file*); + + // Offset of the Additional PLT (if using -z bndplt). + unsigned int aplt_offset_; + + // The size of an entry in the PLT. + static const int plt_entry_size = 16; + + // The size of an entry in the additional PLT. + static const int aplt_entry_size = 8; + + // The first entry in the PLT. + // From the AMD64 ABI: "Unlike Intel386 ABI, this ABI uses the same + // procedure linkage table for both programs and shared objects." + static const unsigned char first_plt_entry[plt_entry_size]; + + // Other entries in the PLT for an executable. + static const unsigned char plt_entry[plt_entry_size]; + + // Entries in the additional PLT. + static const unsigned char aplt_entry[aplt_entry_size]; + + // The reserved TLSDESC entry in the PLT for an executable. + static const unsigned char tlsdesc_plt_entry[plt_entry_size]; + + // The .eh_frame unwind information for the PLT. + static const int plt_eh_frame_fde_size = 32; + static const unsigned char plt_eh_frame_fde[plt_eh_frame_fde_size]; +}; + +// We use this PLT when Indirect Branch Tracking (IBT) is enabled. + +template +class Output_data_plt_x86_64_ibt : public Output_data_plt_x86_64 +{ + public: + Output_data_plt_x86_64_ibt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative) + : Output_data_plt_x86_64(layout, plt_entry_size, + got, got_plt, got_irelative), + aplt_offset_(0) + { } + + Output_data_plt_x86_64_ibt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative, + unsigned int plt_count) + : Output_data_plt_x86_64(layout, plt_entry_size, + got, got_plt, got_irelative, + plt_count), + aplt_offset_(0) + { } + + protected: + virtual unsigned int + do_get_plt_entry_size() const + { return plt_entry_size; } + + // Return the PLT address to use for a global symbol. + uint64_t + do_address_for_global(const Symbol*); + + // Return the PLT address to use for a local symbol. + uint64_t + do_address_for_local(const Relobj*, unsigned int symndx); + + virtual void + do_add_eh_frame(Layout* layout) + { + layout->add_eh_frame_for_plt(this, + this->plt_eh_frame_cie, + this->plt_eh_frame_cie_size, + plt_eh_frame_fde, + plt_eh_frame_fde_size); + } + + virtual void + do_fill_first_plt_entry(unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_addr, + typename elfcpp::Elf_types::Elf_Addr plt_addr); + + virtual unsigned int + do_fill_plt_entry(unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index); + + virtual void + do_fill_tlsdesc_entry(unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address, + typename elfcpp::Elf_types::Elf_Addr got_base, + unsigned int tlsdesc_got_offset, + unsigned int plt_offset); + + void + fill_aplt_entry(unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index); + + private: + // Set the final size. + void + set_final_data_size(); + + // Write out the BND PLT data. + void + do_write(Output_file*); + + // Offset of the Additional PLT (if using -z bndplt). + unsigned int aplt_offset_; + + // The size of an entry in the PLT. + static const int plt_entry_size = 16; + + // The size of an entry in the additional PLT. + static const int aplt_entry_size = 16; + + // The first entry in the PLT. + // From the AMD64 ABI: "Unlike Intel386 ABI, this ABI uses the same + // procedure linkage table for both programs and shared objects." + static const unsigned char first_plt_entry[plt_entry_size]; + + // Other entries in the PLT for an executable. + static const unsigned char plt_entry[plt_entry_size]; + + // Entries in the additional PLT. + static const unsigned char aplt_entry[aplt_entry_size]; + + // The reserved TLSDESC entry in the PLT for an executable. + static const unsigned char tlsdesc_plt_entry[plt_entry_size]; + + // The .eh_frame unwind information for the PLT. + static const int plt_eh_frame_fde_size = 32; + static const unsigned char plt_eh_frame_fde[plt_eh_frame_fde_size]; +}; + +template +class Lazy_view +{ + public: + Lazy_view(Sized_relobj_file* object, unsigned int data_shndx) + : object_(object), data_shndx_(data_shndx), view_(NULL), view_size_(0) + { } + + inline unsigned char + operator[](size_t offset) + { + if (this->view_ == NULL) + this->view_ = this->object_->section_contents(this->data_shndx_, + &this->view_size_, + true); + if (offset >= this->view_size_) + return 0; + return this->view_[offset]; + } + + private: + Sized_relobj_file* object_; + unsigned int data_shndx_; + const unsigned char* view_; + section_size_type view_size_; +}; + // The x86_64 target class. // See the ABI at // http://www.x86-64.org/documentation/abi.pdf @@ -424,7 +705,9 @@ class Target_x86_64 : public Sized_target got_tlsdesc_(NULL), global_offset_table_(NULL), rela_dyn_(NULL), rela_irelative_(NULL), copy_relocs_(elfcpp::R_X86_64_COPY), got_mod_index_offset_(-1U), tlsdesc_reloc_info_(), - tls_base_symbol_defined_(false) + tls_base_symbol_defined_(false), isa_1_used_(0), isa_1_needed_(0), + feature_1_(0), object_isa_1_used_(0), object_feature_1_(0), + seen_first_object_(false) { } // Hook for a new output section. @@ -496,6 +779,21 @@ class Target_x86_64 : public Sized_target const unsigned char* plocal_symbols, Relocatable_relocs*); + // Scan the relocs for --emit-relocs. + void + emit_relocs_scan(Symbol_table* symtab, + Layout* layout, + Sized_relobj_file* object, + unsigned int data_shndx, + unsigned int sh_type, + const unsigned char* prelocs, + size_t reloc_count, + Output_section* output_section, + bool needs_special_offset_handling, + size_t local_symbol_count, + const unsigned char* plocal_syms, + Relocatable_relocs* rr); + // Emit relocations for a section. void relocate_relocs( @@ -505,7 +803,6 @@ class Target_x86_64 : public Sized_target size_t reloc_count, Output_section* output_section, typename elfcpp::Elf_types::Elf_Off offset_in_output_section, - const Relocatable_relocs*, unsigned char* view, typename elfcpp::Elf_types::Elf_Addr view_address, section_size_type view_size, @@ -549,10 +846,13 @@ class Target_x86_64 : public Sized_target // and global_reloc_may_be_function_pointer) // if a function's pointer is taken. ICF uses this in safe mode to only // fold those functions whose pointer is defintely not taken. For x86_64 - // pie binaries, safe ICF cannot be done by looking at relocation types. + // pie binaries, safe ICF cannot be done by looking at only relocation + // types, and for certain cases (e.g. R_X86_64_PC32), the instruction + // opcode is checked as well to distinguish a function call from taking + // a function's pointer. bool do_can_check_for_function_pointers() const - { return !parameters->options().pie(); } + { return true; } // Return the base for a DW_EH_PE_datarel encoding. uint64_t @@ -562,6 +862,7 @@ class Target_x86_64 : public Sized_target void do_calls_non_split(Relobj* object, unsigned int shndx, section_offset_type fnoffset, section_size_type fnsize, + const unsigned char* prelocs, size_t reloc_count, unsigned char* view, section_size_type view_size, std::string* from, std::string* to) const; @@ -594,6 +895,11 @@ class Target_x86_64 : public Sized_target unsigned int plt_entry_size() const; + // Return the size of each GOT entry. + unsigned int + got_entry_size() const + { return 8; }; + // Create the GOT section for an incremental update. Output_data_got_base* init_got_plt_for_update(Symbol_table* symtab, @@ -667,23 +973,14 @@ class Target_x86_64 : public Sized_target do_make_data_plt(Layout* layout, Output_data_got<64, false>* got, Output_data_got_plt_x86_64* got_plt, - Output_data_space* got_irelative) - { - return new Output_data_plt_x86_64_standard(layout, got, got_plt, - got_irelative); - } + Output_data_space* got_irelative); virtual Output_data_plt_x86_64* do_make_data_plt(Layout* layout, Output_data_got<64, false>* got, Output_data_got_plt_x86_64* got_plt, Output_data_space* got_irelative, - unsigned int plt_count) - { - return new Output_data_plt_x86_64_standard(layout, got, got_plt, - got_irelative, - plt_count); - } + unsigned int plt_count); private: // The class which scans relocations. @@ -747,7 +1044,10 @@ class Target_x86_64 : public Sized_target check_non_pic(Relobj*, unsigned int r_type, Symbol*); inline bool - possible_function_pointer_reloc(unsigned int r_type); + possible_function_pointer_reloc(Sized_relobj_file* src_obj, + unsigned int src_indx, + unsigned int r_offset, + unsigned int r_type); bool reloc_needs_plt_for_ifunc(Sized_relobj_file*, @@ -777,11 +1077,9 @@ class Target_x86_64 : public Sized_target // Do a relocation. Return false if the caller should not issue // any warnings about this relocation. inline bool - relocate(const Relocate_info*, Target_x86_64*, - Output_section*, - size_t relnum, const elfcpp::Rela&, - unsigned int r_type, const Sized_symbol*, - const Symbol_value*, + relocate(const Relocate_info*, unsigned int, + Target_x86_64*, Output_section*, size_t, const unsigned char*, + const Sized_symbol*, const Symbol_value*, unsigned char*, typename elfcpp::Elf_types::Elf_Addr, section_size_type); @@ -798,7 +1096,6 @@ class Target_x86_64 : public Sized_target // Do a TLS General-Dynamic to Initial-Exec transition. inline void tls_gd_to_ie(const Relocate_info*, size_t relnum, - Output_segment* tls_segment, const elfcpp::Rela&, unsigned int r_type, typename elfcpp::Elf_types::Elf_Addr value, unsigned char* view, @@ -817,7 +1114,6 @@ class Target_x86_64 : public Sized_target // Do a TLSDESC-style General-Dynamic to Initial-Exec transition. inline void tls_desc_gd_to_ie(const Relocate_info*, size_t relnum, - Output_segment* tls_segment, const elfcpp::Rela&, unsigned int r_type, typename elfcpp::Elf_types::Elf_Addr value, unsigned char* view, @@ -856,14 +1152,70 @@ class Target_x86_64 : public Sized_target bool skip_call_tls_get_addr_; }; - // A class which returns the size required for a relocation type, - // used while scanning relocs during a relocatable link. - class Relocatable_size_for_reloc + // Check if relocation against this symbol is a candidate for + // conversion from + // mov foo@GOTPCREL(%rip), %reg + // to lea foo(%rip), %reg. + template + static inline bool + can_convert_mov_to_lea(const Symbol* gsym, unsigned int r_type, + size_t r_offset, View_type* view) { - public: - unsigned int - get_size_for_reloc(unsigned int, Relobj*); - }; + gold_assert(gsym != NULL); + // We cannot do the conversion unless it's one of these relocations. + if (r_type != elfcpp::R_X86_64_GOTPCREL + && r_type != elfcpp::R_X86_64_GOTPCRELX + && r_type != elfcpp::R_X86_64_REX_GOTPCRELX) + return false; + // We cannot convert references to IFUNC symbols, or to symbols that + // are not local to the current module. + // We can't do predefined symbols because they may become undefined + // (e.g., __ehdr_start when the headers aren't mapped to a segment). + if (gsym->type() == elfcpp::STT_GNU_IFUNC + || gsym->is_undefined() + || gsym->is_predefined() + || gsym->is_from_dynobj() + || gsym->is_preemptible()) + return false; + // If we are building a shared object and the symbol is protected, we may + // need to go through the GOT. + if (parameters->options().shared() + && gsym->visibility() == elfcpp::STV_PROTECTED) + return false; + // We cannot convert references to the _DYNAMIC symbol. + if (strcmp(gsym->name(), "_DYNAMIC") == 0) + return false; + // Check for a MOV opcode. + return (*view)[r_offset - 2] == 0x8b; + } + + // Convert + // callq *foo@GOTPCRELX(%rip) to + // addr32 callq foo + // and jmpq *foo@GOTPCRELX(%rip) to + // jmpq foo + // nop + template + static inline bool + can_convert_callq_to_direct(const Symbol* gsym, unsigned int r_type, + size_t r_offset, View_type* view) + { + gold_assert(gsym != NULL); + // We cannot do the conversion unless it's a GOTPCRELX relocation. + if (r_type != elfcpp::R_X86_64_GOTPCRELX) + return false; + // We cannot convert references to IFUNC symbols, or to symbols that + // are not local to the current module. + if (gsym->type() == elfcpp::STT_GNU_IFUNC + || gsym->is_undefined () + || gsym->is_from_dynobj() + || gsym->is_preemptible()) + return false; + // Check for a CALLQ or JMPQ opcode. + return ((*view)[r_offset - 2] == 0xff + && ((*view)[r_offset - 1] == 0x15 + || (*view)[r_offset - 1] == 0x25)); + } // Adjust TLS relocation type based on the options and whether this // is a local symbol. @@ -944,12 +1296,30 @@ class Target_x86_64 : public Sized_target unsigned int shndx, Output_section* output_section, Symbol* sym, const elfcpp::Rela& reloc) { + unsigned int r_type = elfcpp::elf_r_type(reloc.get_r_info()); this->copy_relocs_.copy_reloc(symtab, layout, symtab->get_sized_symbol(sym), object, shndx, output_section, - reloc, this->rela_dyn_section(layout)); + r_type, reloc.get_r_offset(), + reloc.get_r_addend(), + this->rela_dyn_section(layout)); } + // Record a target-specific program property in the .note.gnu.property + // section. + void + record_gnu_property(unsigned int, unsigned int, size_t, + const unsigned char*, const Object*); + + // Merge the target-specific program properties from the current object. + void + merge_gnu_properties(const Object*); + + // Finalize the target-specific program properties and add them back to + // the layout. + void + do_finalize_gnu_properties(Layout*) const; + // Information about this specific target which we pass to the // general Target structure. static const Target::Target_info x86_64_info; @@ -1007,6 +1377,22 @@ class Target_x86_64 : public Sized_target std::vector tlsdesc_reloc_info_; // True if the _TLS_MODULE_BASE_ symbol has been defined. bool tls_base_symbol_defined_; + // Target-specific program properties, from .note.gnu.property section. + // Each bit represents a specific feature. + uint32_t isa_1_used_; + uint32_t isa_1_needed_; + uint32_t feature_1_; + // Target-specific properties from the current object. + // These bits get ORed into ISA_1_USED_ after all properties for the object + // have been processed. But if either is all zeroes (as when the property + // is absent from an object), the result should be all zeroes. + // (See PR ld/23486.) + uint32_t object_isa_1_used_; + // These bits get ANDed into FEATURE_1_ after all properties for the object + // have been processed. + uint32_t object_feature_1_; + // Whether we have seen our first object, for use in initializing FEATURE_1_. + bool seen_first_object_; }; template<> @@ -1033,7 +1419,9 @@ const Target::Target_info Target_x86_64<64>::x86_64_info = elfcpp::SHF_X86_64_LARGE, // large_common_section_flags NULL, // attributes_section NULL, // attributes_vendor - "_start" // entry_symbol_name + "_start", // entry_symbol_name + 32, // hash_entry_size + elfcpp::SHT_X86_64_UNWIND, // unwind_section_type }; template<> @@ -1060,7 +1448,9 @@ const Target::Target_info Target_x86_64<32>::x86_64_info = elfcpp::SHF_X86_64_LARGE, // large_common_section_flags NULL, // attributes_section NULL, // attributes_vendor - "_start" // entry_symbol_name + "_start", // entry_symbol_name + 32, // hash_entry_size + elfcpp::SHT_X86_64_UNWIND, // unwind_section_type }; // This is called when a new output section is created. This is where @@ -1191,6 +1581,103 @@ Target_x86_64::rela_irelative_section(Layout* layout) return this->rela_irelative_; } +// Record a target-specific program property from the .note.gnu.property +// section. +template +void +Target_x86_64::record_gnu_property( + unsigned int, unsigned int pr_type, + size_t pr_datasz, const unsigned char* pr_data, + const Object* object) +{ + uint32_t val = 0; + + switch (pr_type) + { + case elfcpp::GNU_PROPERTY_X86_ISA_1_USED: + case elfcpp::GNU_PROPERTY_X86_ISA_1_NEEDED: + case elfcpp::GNU_PROPERTY_X86_FEATURE_1_AND: + if (pr_datasz != 4) + { + gold_warning(_("%s: corrupt .note.gnu.property section " + "(pr_datasz for property %d is not 4)"), + object->name().c_str(), pr_type); + return; + } + val = elfcpp::Swap<32, false>::readval(pr_data); + break; + default: + gold_warning(_("%s: unknown program property type 0x%x " + "in .note.gnu.property section"), + object->name().c_str(), pr_type); + break; + } + + switch (pr_type) + { + case elfcpp::GNU_PROPERTY_X86_ISA_1_USED: + this->object_isa_1_used_ |= val; + break; + case elfcpp::GNU_PROPERTY_X86_ISA_1_NEEDED: + this->isa_1_needed_ |= val; + break; + case elfcpp::GNU_PROPERTY_X86_FEATURE_1_AND: + // If we see multiple feature props in one object, OR them together. + this->object_feature_1_ |= val; + break; + } +} + +// Merge the target-specific program properties from the current object. +template +void +Target_x86_64::merge_gnu_properties(const Object*) +{ + if (this->seen_first_object_) + { + // If any object is missing the ISA_1_USED property, we must omit + // it from the output file. + if (this->object_isa_1_used_ == 0) + this->isa_1_used_ = 0; + else if (this->isa_1_used_ != 0) + this->isa_1_used_ |= this->object_isa_1_used_; + this->feature_1_ &= this->object_feature_1_; + } + else + { + this->isa_1_used_ = this->object_isa_1_used_; + this->feature_1_ = this->object_feature_1_; + this->seen_first_object_ = true; + } + this->object_isa_1_used_ = 0; + this->object_feature_1_ = 0; +} + +static inline void +add_property(Layout* layout, unsigned int pr_type, uint32_t val) +{ + unsigned char buf[4]; + elfcpp::Swap<32, false>::writeval(buf, val); + layout->add_gnu_property(elfcpp::NT_GNU_PROPERTY_TYPE_0, pr_type, 4, buf); +} + +// Finalize the target-specific program properties and add them back to +// the layout. +template +void +Target_x86_64::do_finalize_gnu_properties(Layout* layout) const +{ + if (this->isa_1_used_ != 0) + add_property(layout, elfcpp::GNU_PROPERTY_X86_ISA_1_USED, + this->isa_1_used_); + if (this->isa_1_needed_ != 0) + add_property(layout, elfcpp::GNU_PROPERTY_X86_ISA_1_NEEDED, + this->isa_1_needed_); + if (this->feature_1_ != 0) + add_property(layout, elfcpp::GNU_PROPERTY_X86_FEATURE_1_AND, + this->feature_1_); +} + // Write the first three reserved words of the .got.plt section. // The remainder of the section is written while writing the PLT // in Output_data_plt_i386::do_write. @@ -1427,7 +1914,7 @@ Output_data_plt_x86_64::rela_irelative(Symbol_table* symtab, template uint64_t -Output_data_plt_x86_64::address_for_global(const Symbol* gsym) +Output_data_plt_x86_64::do_address_for_global(const Symbol* gsym) { uint64_t offset = 0; if (gsym->type() == elfcpp::STT_GNU_IFUNC @@ -1441,8 +1928,8 @@ Output_data_plt_x86_64::address_for_global(const Symbol* gsym) template uint64_t -Output_data_plt_x86_64::address_for_local(const Relobj* object, - unsigned int r_sym) +Output_data_plt_x86_64::do_address_for_local(const Relobj* object, + unsigned int r_sym) { return (this->address() + (this->count_ + 1) * this->get_plt_entry_size() @@ -1454,10 +1941,12 @@ template void Output_data_plt_x86_64::set_final_data_size() { - unsigned int count = this->count_ + this->irelative_count_; + // Number of regular and IFUNC PLT entries, plus the first entry. + unsigned int count = this->count_ + this->irelative_count_ + 1; + // Count the TLSDESC entry, if present. if (this->has_tlsdesc_entry()) ++count; - this->set_data_size((count + 1) * this->get_plt_entry_size()); + this->set_data_size(count * this->get_plt_entry_size()); } // The first entry in the PLT for an executable. @@ -1516,11 +2005,16 @@ Output_data_plt_x86_64_standard::do_fill_plt_entry( unsigned int plt_offset, unsigned int plt_index) { + // Check PC-relative offset overflow in PLT entry. + uint64_t plt_got_pcrel_offset = (got_address + got_offset + - (plt_address + plt_offset + 6)); + if (Bits<32>::has_overflow(plt_got_pcrel_offset)) + gold_error(_("PC-relative offset overflow in PLT entry %d"), + plt_index + 1); + memcpy(pov, plt_entry, plt_entry_size); elfcpp::Swap_unaligned<32, false>::writeval(pov + 2, - (got_address + got_offset - - (plt_address + plt_offset - + 6))); + plt_got_pcrel_offset); elfcpp::Swap_unaligned<32, false>::writeval(pov + 7, plt_index); elfcpp::Swap<32, false>::writeval(pov + 12, @@ -1567,11 +2061,419 @@ Output_data_plt_x86_64_standard::do_fill_tlsdesc_entry( + 12))); } -// The .eh_frame unwind information for the PLT. +// Return the APLT address to use for a global symbol (for -z bndplt). -template -const unsigned char -Output_data_plt_x86_64::plt_eh_frame_cie[plt_eh_frame_cie_size] = +uint64_t +Output_data_plt_x86_64_bnd::do_address_for_global(const Symbol* gsym) +{ + uint64_t offset = this->aplt_offset_; + // Convert the PLT offset into an APLT offset. + unsigned int plt_offset = gsym->plt_offset(); + if (gsym->type() == elfcpp::STT_GNU_IFUNC + && gsym->can_use_relative_reloc(false)) + offset += this->regular_count() * aplt_entry_size; + else + plt_offset -= plt_entry_size; + plt_offset = plt_offset / (plt_entry_size / aplt_entry_size); + return this->address() + offset + plt_offset; +} + +// Return the PLT address to use for a local symbol. These are always +// IRELATIVE relocs. + +uint64_t +Output_data_plt_x86_64_bnd::do_address_for_local(const Relobj* object, + unsigned int r_sym) +{ + // Convert the PLT offset into an APLT offset. + unsigned int plt_offset = ((object->local_plt_offset(r_sym) - plt_entry_size) + / (plt_entry_size / aplt_entry_size)); + return (this->address() + + this->aplt_offset_ + + this->regular_count() * aplt_entry_size + + plt_offset); +} + +// Set the final size. +void +Output_data_plt_x86_64_bnd::set_final_data_size() +{ + // Number of regular and IFUNC PLT entries. + unsigned int count = this->entry_count(); + // Count the first entry and the TLSDESC entry, if present. + unsigned int extra = this->has_tlsdesc_entry() ? 2 : 1; + unsigned int plt_size = (count + extra) * plt_entry_size; + // Offset of the APLT. + this->aplt_offset_ = plt_size; + // Size of the APLT. + plt_size += count * aplt_entry_size; + this->set_data_size(plt_size); +} + +// The first entry in the BND PLT. + +const unsigned char +Output_data_plt_x86_64_bnd::first_plt_entry[plt_entry_size] = +{ + // From AMD64 ABI Draft 0.98, page 76 + 0xff, 0x35, // pushq contents of memory address + 0, 0, 0, 0, // replaced with address of .got + 8 + 0xf2, 0xff, 0x25, // bnd jmp indirect + 0, 0, 0, 0, // replaced with address of .got + 16 + 0x0f, 0x1f, 0x00 // nop +}; + +void +Output_data_plt_x86_64_bnd::do_fill_first_plt_entry( + unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address) +{ + memcpy(pov, first_plt_entry, plt_entry_size); + // We do a jmp relative to the PC at the end of this instruction. + elfcpp::Swap_unaligned<32, false>::writeval(pov + 2, + (got_address + 8 + - (plt_address + 6))); + elfcpp::Swap<32, false>::writeval(pov + 9, + (got_address + 16 + - (plt_address + 13))); +} + +// Subsequent entries in the BND PLT. + +const unsigned char +Output_data_plt_x86_64_bnd::plt_entry[plt_entry_size] = +{ + // From AMD64 ABI Draft 0.99.8, page 139 + 0x68, // pushq immediate + 0, 0, 0, 0, // replaced with offset into relocation table + 0xf2, 0xe9, // bnd jmpq relative + 0, 0, 0, 0, // replaced with offset to start of .plt + 0x0f, 0x1f, 0x44, 0, 0 // nop +}; + +// Entries in the BND Additional PLT. + +const unsigned char +Output_data_plt_x86_64_bnd::aplt_entry[aplt_entry_size] = +{ + // From AMD64 ABI Draft 0.99.8, page 139 + 0xf2, 0xff, 0x25, // bnd jmpq indirect + 0, 0, 0, 0, // replaced with address of symbol in .got + 0x90, // nop +}; + +unsigned int +Output_data_plt_x86_64_bnd::do_fill_plt_entry( + unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr, + elfcpp::Elf_types<64>::Elf_Addr, + unsigned int, + unsigned int plt_offset, + unsigned int plt_index) +{ + memcpy(pov, plt_entry, plt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 1, plt_index); + elfcpp::Swap<32, false>::writeval(pov + 7, -(plt_offset + 11)); + return 0; +} + +void +Output_data_plt_x86_64_bnd::fill_aplt_entry( + unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index) +{ + // Check PC-relative offset overflow in PLT entry. + uint64_t plt_got_pcrel_offset = (got_address + got_offset + - (plt_address + plt_offset + 7)); + if (Bits<32>::has_overflow(plt_got_pcrel_offset)) + gold_error(_("PC-relative offset overflow in APLT entry %d"), + plt_index + 1); + + memcpy(pov, aplt_entry, aplt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 3, plt_got_pcrel_offset); +} + +// The reserved TLSDESC entry in the PLT for an executable. + +const unsigned char +Output_data_plt_x86_64_bnd::tlsdesc_plt_entry[plt_entry_size] = +{ + // From Alexandre Oliva, "Thread-Local Storage Descriptors for IA32 + // and AMD64/EM64T", Version 0.9.4 (2005-10-10). + 0xff, 0x35, // pushq x(%rip) + 0, 0, 0, 0, // replaced with address of linkmap GOT entry (at PLTGOT + 8) + 0xf2, 0xff, 0x25, // jmpq *y(%rip) + 0, 0, 0, 0, // replaced with offset of reserved TLSDESC_GOT entry + 0x0f, 0x1f, 0 // nop +}; + +void +Output_data_plt_x86_64_bnd::do_fill_tlsdesc_entry( + unsigned char* pov, + elfcpp::Elf_types<64>::Elf_Addr got_address, + elfcpp::Elf_types<64>::Elf_Addr plt_address, + elfcpp::Elf_types<64>::Elf_Addr got_base, + unsigned int tlsdesc_got_offset, + unsigned int plt_offset) +{ + memcpy(pov, tlsdesc_plt_entry, plt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 2, + (got_address + 8 + - (plt_address + plt_offset + + 6))); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 9, + (got_base + + tlsdesc_got_offset + - (plt_address + plt_offset + + 13))); +} + +// Return the APLT address to use for a global symbol (for IBT). + +template +uint64_t +Output_data_plt_x86_64_ibt::do_address_for_global(const Symbol* gsym) +{ + uint64_t offset = this->aplt_offset_; + // Convert the PLT offset into an APLT offset. + unsigned int plt_offset = gsym->plt_offset(); + if (gsym->type() == elfcpp::STT_GNU_IFUNC + && gsym->can_use_relative_reloc(false)) + offset += this->regular_count() * aplt_entry_size; + else + plt_offset -= plt_entry_size; + plt_offset = plt_offset / (plt_entry_size / aplt_entry_size); + return this->address() + offset + plt_offset; +} + +// Return the PLT address to use for a local symbol. These are always +// IRELATIVE relocs. + +template +uint64_t +Output_data_plt_x86_64_ibt::do_address_for_local(const Relobj* object, + unsigned int r_sym) +{ + // Convert the PLT offset into an APLT offset. + unsigned int plt_offset = ((object->local_plt_offset(r_sym) - plt_entry_size) + / (plt_entry_size / aplt_entry_size)); + return (this->address() + + this->aplt_offset_ + + this->regular_count() * aplt_entry_size + + plt_offset); +} + +// Set the final size. + +template +void +Output_data_plt_x86_64_ibt::set_final_data_size() +{ + // Number of regular and IFUNC PLT entries. + unsigned int count = this->entry_count(); + // Count the first entry and the TLSDESC entry, if present. + unsigned int extra = this->has_tlsdesc_entry() ? 2 : 1; + unsigned int plt_size = (count + extra) * plt_entry_size; + // Offset of the APLT. + this->aplt_offset_ = plt_size; + // Size of the APLT. + plt_size += count * aplt_entry_size; + this->set_data_size(plt_size); +} + +// The first entry in the IBT PLT. + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<32>::first_plt_entry[plt_entry_size] = +{ + // MPX isn't supported for x32, so we don't need the BND prefix. + // From AMD64 ABI Draft 0.98, page 76 + 0xff, 0x35, // pushq contents of memory address + 0, 0, 0, 0, // replaced with address of .got + 8 + 0xff, 0x25, // jmp indirect + 0, 0, 0, 0, // replaced with address of .got + 16 + 0x90, 0x90, 0x90, 0x90 // noop (x4) +}; + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<64>::first_plt_entry[plt_entry_size] = +{ + // Use the BND prefix so that IBT is compatible with MPX. + 0xff, 0x35, // pushq contents of memory address + 0, 0, 0, 0, // replaced with address of .got + 8 + 0xf2, 0xff, 0x25, // bnd jmp indirect + 0, 0, 0, 0, // replaced with address of .got + 16 + 0x0f, 0x1f, 0x00 // nop +}; + +template +void +Output_data_plt_x86_64_ibt::do_fill_first_plt_entry( + unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address) +{ + // Offsets to the addresses needing relocation. + const unsigned int roff1 = 2; + const unsigned int roff2 = (size == 32) ? 8 : 9; + + memcpy(pov, first_plt_entry, plt_entry_size); + // We do a jmp relative to the PC at the end of this instruction. + elfcpp::Swap_unaligned<32, false>::writeval(pov + roff1, + (got_address + 8 + - (plt_address + roff1 + 4))); + elfcpp::Swap<32, false>::writeval(pov + roff2, + (got_address + 16 + - (plt_address + roff2 + 4))); +} + +// Subsequent entries in the IBT PLT. + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<32>::plt_entry[plt_entry_size] = +{ + // From AMD64 ABI Draft 1.0-rc1, Chapter 13. + 0xf3, 0x0f, 0x1e, 0xfa, // endbr64 + 0x68, // pushq immediate + 0, 0, 0, 0, // replaced with offset into relocation table + 0xe9, // jmpq relative + 0, 0, 0, 0, // replaced with offset to start of .plt + 0x90, 0x90 // nop +}; + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<64>::plt_entry[plt_entry_size] = +{ + // From AMD64 ABI Draft 1.0-rc1, Chapter 13. + 0xf3, 0x0f, 0x1e, 0xfa, // endbr64 + 0x68, // pushq immediate + 0, 0, 0, 0, // replaced with offset into relocation table + 0xf2, 0xe9, // bnd jmpq relative + 0, 0, 0, 0, // replaced with offset to start of .plt + 0x90 // nop +}; + +// Entries in the IBT Additional PLT. + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<32>::aplt_entry[aplt_entry_size] = +{ + // From AMD64 ABI Draft 1.0-rc1, Chapter 13. + 0xf3, 0x0f, 0x1e, 0xfa, // endbr64 + 0xff, 0x25, // jmpq indirect + 0, 0, 0, 0, // replaced with address of symbol in .got + 0x0f, 0x1f, 0x04, 0x00, // nop + 0x90, 0x90 // nop +}; + +template<> +const unsigned char +Output_data_plt_x86_64_ibt<64>::aplt_entry[aplt_entry_size] = +{ + // From AMD64 ABI Draft 1.0-rc1, Chapter 13. + 0xf3, 0x0f, 0x1e, 0xfa, // endbr64 + 0xf2, 0xff, 0x25, // bnd jmpq indirect + 0, 0, 0, 0, // replaced with address of symbol in .got + 0x0f, 0x1f, 0x04, 0x00, // nop + 0x90, // nop +}; + +template +unsigned int +Output_data_plt_x86_64_ibt::do_fill_plt_entry( + unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr, + typename elfcpp::Elf_types::Elf_Addr, + unsigned int, + unsigned int plt_offset, + unsigned int plt_index) +{ + // Offsets to the addresses needing relocation. + const unsigned int roff1 = 5; + const unsigned int roff2 = (size == 32) ? 10 : 11; + + memcpy(pov, plt_entry, plt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + roff1, plt_index); + elfcpp::Swap<32, false>::writeval(pov + roff2, -(plt_offset + roff2 + 4)); + return 0; +} + +template +void +Output_data_plt_x86_64_ibt::fill_aplt_entry( + unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address, + unsigned int got_offset, + unsigned int plt_offset, + unsigned int plt_index) +{ + // Offset to the address needing relocation. + const unsigned int roff = (size == 32) ? 6 : 7; + + // Check PC-relative offset overflow in PLT entry. + uint64_t plt_got_pcrel_offset = (got_address + got_offset + - (plt_address + plt_offset + roff + 4)); + if (Bits<32>::has_overflow(plt_got_pcrel_offset)) + gold_error(_("PC-relative offset overflow in APLT entry %d"), + plt_index + 1); + + memcpy(pov, aplt_entry, aplt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + roff, plt_got_pcrel_offset); +} + +// The reserved TLSDESC entry in the IBT PLT for an executable. + +template +const unsigned char +Output_data_plt_x86_64_ibt::tlsdesc_plt_entry[plt_entry_size] = +{ + // From Alexandre Oliva, "Thread-Local Storage Descriptors for IA32 + // and AMD64/EM64T", Version 0.9.4 (2005-10-10). + 0xff, 0x35, // pushq x(%rip) + 0, 0, 0, 0, // replaced with address of linkmap GOT entry (at PLTGOT + 8) + 0xf2, 0xff, 0x25, // jmpq *y(%rip) + 0, 0, 0, 0, // replaced with offset of reserved TLSDESC_GOT entry + 0x0f, 0x1f, 0 // nop +}; + +template +void +Output_data_plt_x86_64_ibt::do_fill_tlsdesc_entry( + unsigned char* pov, + typename elfcpp::Elf_types::Elf_Addr got_address, + typename elfcpp::Elf_types::Elf_Addr plt_address, + typename elfcpp::Elf_types::Elf_Addr got_base, + unsigned int tlsdesc_got_offset, + unsigned int plt_offset) +{ + memcpy(pov, tlsdesc_plt_entry, plt_entry_size); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 2, + (got_address + 8 + - (plt_address + plt_offset + + 6))); + elfcpp::Swap_unaligned<32, false>::writeval(pov + 9, + (got_base + + tlsdesc_got_offset + - (plt_address + plt_offset + + 13))); +} + +// The .eh_frame unwind information for the PLT. + +template +const unsigned char +Output_data_plt_x86_64::plt_eh_frame_cie[plt_eh_frame_cie_size] = { 1, // CIE version. 'z', // Augmentation: augmentation size included. @@ -1617,56 +2519,288 @@ Output_data_plt_x86_64_standard::plt_eh_frame_fde[plt_eh_frame_fde_size] = elfcpp::DW_CFA_nop }; -// Write out the PLT. This uses the hand-coded instructions above, -// and adjusts them as needed. This is specified by the AMD64 ABI. +// The .eh_frame unwind information for the BND PLT. +const unsigned char +Output_data_plt_x86_64_bnd::plt_eh_frame_fde[plt_eh_frame_fde_size] = +{ + 0, 0, 0, 0, // Replaced with offset to .plt. + 0, 0, 0, 0, // Replaced with size of .plt. + 0, // Augmentation size. + elfcpp::DW_CFA_def_cfa_offset, 16, // DW_CFA_def_cfa_offset: 16. + elfcpp::DW_CFA_advance_loc + 6, // Advance 6 to __PLT__ + 6. + elfcpp::DW_CFA_def_cfa_offset, 24, // DW_CFA_def_cfa_offset: 24. + elfcpp::DW_CFA_advance_loc + 10, // Advance 10 to __PLT__ + 16. + elfcpp::DW_CFA_def_cfa_expression, // DW_CFA_def_cfa_expression. + 11, // Block length. + elfcpp::DW_OP_breg7, 8, // Push %rsp + 8. + elfcpp::DW_OP_breg16, 0, // Push %rip. + elfcpp::DW_OP_lit15, // Push 0xf. + elfcpp::DW_OP_and, // & (%rip & 0xf). + elfcpp::DW_OP_lit5, // Push 5. + elfcpp::DW_OP_ge, // >= ((%rip & 0xf) >= 5) + elfcpp::DW_OP_lit3, // Push 3. + elfcpp::DW_OP_shl, // << (((%rip & 0xf) >= 5) << 3) + elfcpp::DW_OP_plus, // + ((((%rip&0xf)>=5)<<3)+%rsp+8 + elfcpp::DW_CFA_nop, // Align to 32 bytes. + elfcpp::DW_CFA_nop, + elfcpp::DW_CFA_nop, + elfcpp::DW_CFA_nop +}; + +// The .eh_frame unwind information for the BND PLT. +template +const unsigned char +Output_data_plt_x86_64_ibt::plt_eh_frame_fde[plt_eh_frame_fde_size] = +{ + 0, 0, 0, 0, // Replaced with offset to .plt. + 0, 0, 0, 0, // Replaced with size of .plt. + 0, // Augmentation size. + elfcpp::DW_CFA_def_cfa_offset, 16, // DW_CFA_def_cfa_offset: 16. + elfcpp::DW_CFA_advance_loc + 6, // Advance 6 to __PLT__ + 6. + elfcpp::DW_CFA_def_cfa_offset, 24, // DW_CFA_def_cfa_offset: 24. + elfcpp::DW_CFA_advance_loc + 10, // Advance 10 to __PLT__ + 16. + elfcpp::DW_CFA_def_cfa_expression, // DW_CFA_def_cfa_expression. + 11, // Block length. + elfcpp::DW_OP_breg7, 8, // Push %rsp + 8. + elfcpp::DW_OP_breg16, 0, // Push %rip. + elfcpp::DW_OP_lit15, // Push 0xf. + elfcpp::DW_OP_and, // & (%rip & 0xf). + elfcpp::DW_OP_lit9, // Push 9. + elfcpp::DW_OP_ge, // >= ((%rip & 0xf) >= 9) + elfcpp::DW_OP_lit3, // Push 3. + elfcpp::DW_OP_shl, // << (((%rip & 0xf) >= 9) << 3) + elfcpp::DW_OP_plus, // + ((((%rip&0xf)>=9)<<3)+%rsp+8 + elfcpp::DW_CFA_nop, // Align to 32 bytes. + elfcpp::DW_CFA_nop, + elfcpp::DW_CFA_nop, + elfcpp::DW_CFA_nop +}; + +// Write out the PLT. This uses the hand-coded instructions above, +// and adjusts them as needed. This is specified by the AMD64 ABI. + +template +void +Output_data_plt_x86_64::do_write(Output_file* of) +{ + const off_t offset = this->offset(); + const section_size_type oview_size = + convert_to_section_size_type(this->data_size()); + unsigned char* const oview = of->get_output_view(offset, oview_size); + + const off_t got_file_offset = this->got_plt_->offset(); + gold_assert(parameters->incremental_update() + || (got_file_offset + this->got_plt_->data_size() + == this->got_irelative_->offset())); + const section_size_type got_size = + convert_to_section_size_type(this->got_plt_->data_size() + + this->got_irelative_->data_size()); + unsigned char* const got_view = of->get_output_view(got_file_offset, + got_size); + + unsigned char* pov = oview; + + // The base address of the .plt section. + typename elfcpp::Elf_types::Elf_Addr plt_address = this->address(); + // The base address of the .got section. + typename elfcpp::Elf_types::Elf_Addr got_base = this->got_->address(); + // The base address of the PLT portion of the .got section, + // which is where the GOT pointer will point, and where the + // three reserved GOT entries are located. + typename elfcpp::Elf_types::Elf_Addr got_address + = this->got_plt_->address(); + + this->fill_first_plt_entry(pov, got_address, plt_address); + pov += this->get_plt_entry_size(); + + // The first three entries in the GOT are reserved, and are written + // by Output_data_got_plt_x86_64::do_write. + unsigned char* got_pov = got_view + 24; + + unsigned int plt_offset = this->get_plt_entry_size(); + unsigned int got_offset = 24; + const unsigned int count = this->count_ + this->irelative_count_; + for (unsigned int plt_index = 0; + plt_index < count; + ++plt_index, + pov += this->get_plt_entry_size(), + got_pov += 8, + plt_offset += this->get_plt_entry_size(), + got_offset += 8) + { + // Set and adjust the PLT entry itself. + unsigned int lazy_offset = this->fill_plt_entry(pov, + got_address, plt_address, + got_offset, plt_offset, + plt_index); + + // Set the entry in the GOT. + elfcpp::Swap<64, false>::writeval(got_pov, + plt_address + plt_offset + lazy_offset); + } + + if (this->has_tlsdesc_entry()) + { + // Set and adjust the reserved TLSDESC PLT entry. + unsigned int tlsdesc_got_offset = this->get_tlsdesc_got_offset(); + this->fill_tlsdesc_entry(pov, got_address, plt_address, got_base, + tlsdesc_got_offset, plt_offset); + pov += this->get_plt_entry_size(); + } + + gold_assert(static_cast(pov - oview) == oview_size); + gold_assert(static_cast(got_pov - got_view) == got_size); + + of->write_output_view(offset, oview_size, oview); + of->write_output_view(got_file_offset, got_size, got_view); +} + +// Write out the BND PLT. + +void +Output_data_plt_x86_64_bnd::do_write(Output_file* of) +{ + const off_t offset = this->offset(); + const section_size_type oview_size = + convert_to_section_size_type(this->data_size()); + unsigned char* const oview = of->get_output_view(offset, oview_size); + + Output_data_got<64, false>* got = this->got(); + Output_data_got_plt_x86_64* got_plt = this->got_plt(); + Output_data_space* got_irelative = this->got_irelative(); + + const off_t got_file_offset = got_plt->offset(); + gold_assert(parameters->incremental_update() + || (got_file_offset + got_plt->data_size() + == got_irelative->offset())); + const section_size_type got_size = + convert_to_section_size_type(got_plt->data_size() + + got_irelative->data_size()); + unsigned char* const got_view = of->get_output_view(got_file_offset, + got_size); + + unsigned char* pov = oview; + + // The base address of the .plt section. + elfcpp::Elf_types<64>::Elf_Addr plt_address = this->address(); + // The base address of the .got section. + elfcpp::Elf_types<64>::Elf_Addr got_base = got->address(); + // The base address of the PLT portion of the .got section, + // which is where the GOT pointer will point, and where the + // three reserved GOT entries are located. + elfcpp::Elf_types<64>::Elf_Addr got_address = got_plt->address(); + + this->fill_first_plt_entry(pov, got_address, plt_address); + pov += plt_entry_size; + + // The first three entries in the GOT are reserved, and are written + // by Output_data_got_plt_x86_64::do_write. + unsigned char* got_pov = got_view + 24; + + unsigned int plt_offset = plt_entry_size; + unsigned int got_offset = 24; + const unsigned int count = this->entry_count(); + for (unsigned int plt_index = 0; + plt_index < count; + ++plt_index, + pov += plt_entry_size, + got_pov += 8, + plt_offset += plt_entry_size, + got_offset += 8) + { + // Set and adjust the PLT entry itself. + unsigned int lazy_offset = this->fill_plt_entry(pov, + got_address, plt_address, + got_offset, plt_offset, + plt_index); + + // Set the entry in the GOT. + elfcpp::Swap<64, false>::writeval(got_pov, + plt_address + plt_offset + lazy_offset); + } + + if (this->has_tlsdesc_entry()) + { + // Set and adjust the reserved TLSDESC PLT entry. + unsigned int tlsdesc_got_offset = this->get_tlsdesc_got_offset(); + this->fill_tlsdesc_entry(pov, got_address, plt_address, got_base, + tlsdesc_got_offset, plt_offset); + pov += this->get_plt_entry_size(); + } + + // Write the additional PLT. + got_offset = 24; + for (unsigned int plt_index = 0; + plt_index < count; + ++plt_index, + pov += aplt_entry_size, + plt_offset += aplt_entry_size, + got_offset += 8) + { + // Set and adjust the APLT entry. + this->fill_aplt_entry(pov, got_address, plt_address, got_offset, + plt_offset, plt_index); + } + + gold_assert(static_cast(pov - oview) == oview_size); + gold_assert(static_cast(got_pov - got_view) == got_size); + + of->write_output_view(offset, oview_size, oview); + of->write_output_view(got_file_offset, got_size, got_view); +} + +// Write out the IBT PLT. template void -Output_data_plt_x86_64::do_write(Output_file* of) +Output_data_plt_x86_64_ibt::do_write(Output_file* of) { const off_t offset = this->offset(); const section_size_type oview_size = convert_to_section_size_type(this->data_size()); unsigned char* const oview = of->get_output_view(offset, oview_size); - const off_t got_file_offset = this->got_plt_->offset(); + Output_data_got<64, false>* got = this->got(); + Output_data_got_plt_x86_64* got_plt = this->got_plt(); + Output_data_space* got_irelative = this->got_irelative(); + + const off_t got_file_offset = got_plt->offset(); gold_assert(parameters->incremental_update() - || (got_file_offset + this->got_plt_->data_size() - == this->got_irelative_->offset())); + || (got_file_offset + got_plt->data_size() + == got_irelative->offset())); const section_size_type got_size = - convert_to_section_size_type(this->got_plt_->data_size() - + this->got_irelative_->data_size()); + convert_to_section_size_type(got_plt->data_size() + + got_irelative->data_size()); unsigned char* const got_view = of->get_output_view(got_file_offset, got_size); unsigned char* pov = oview; // The base address of the .plt section. - typename elfcpp::Elf_types::Elf_Addr plt_address = this->address(); + elfcpp::Elf_types<64>::Elf_Addr plt_address = this->address(); // The base address of the .got section. - typename elfcpp::Elf_types::Elf_Addr got_base = this->got_->address(); + elfcpp::Elf_types<64>::Elf_Addr got_base = got->address(); // The base address of the PLT portion of the .got section, // which is where the GOT pointer will point, and where the // three reserved GOT entries are located. - typename elfcpp::Elf_types::Elf_Addr got_address - = this->got_plt_->address(); + elfcpp::Elf_types<64>::Elf_Addr got_address = got_plt->address(); this->fill_first_plt_entry(pov, got_address, plt_address); - pov += this->get_plt_entry_size(); + pov += plt_entry_size; // The first three entries in the GOT are reserved, and are written // by Output_data_got_plt_x86_64::do_write. unsigned char* got_pov = got_view + 24; - unsigned int plt_offset = this->get_plt_entry_size(); + unsigned int plt_offset = plt_entry_size; unsigned int got_offset = 24; - const unsigned int count = this->count_ + this->irelative_count_; + const unsigned int count = this->entry_count(); for (unsigned int plt_index = 0; plt_index < count; ++plt_index, - pov += this->get_plt_entry_size(), + pov += plt_entry_size, got_pov += 8, - plt_offset += this->get_plt_entry_size(), + plt_offset += plt_entry_size, got_offset += 8) { // Set and adjust the PLT entry itself. @@ -1689,6 +2823,20 @@ Output_data_plt_x86_64::do_write(Output_file* of) pov += this->get_plt_entry_size(); } + // Write the additional PLT. + got_offset = 24; + for (unsigned int plt_index = 0; + plt_index < count; + ++plt_index, + pov += aplt_entry_size, + plt_offset += aplt_entry_size, + got_offset += 8) + { + // Set and adjust the APLT entry. + this->fill_aplt_entry(pov, got_address, plt_address, got_offset, + plt_offset, plt_index); + } + gold_assert(static_cast(pov - oview) == oview_size); gold_assert(static_cast(got_pov - got_view) == got_size); @@ -1725,6 +2873,73 @@ Target_x86_64::make_plt_section(Symbol_table* symtab, Layout* layout) } } +template<> +Output_data_plt_x86_64<32>* +Target_x86_64<32>::do_make_data_plt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative) +{ + if (this->feature_1_ & elfcpp::GNU_PROPERTY_X86_FEATURE_1_IBT) + return new Output_data_plt_x86_64_ibt<32>(layout, got, got_plt, + got_irelative); + return new Output_data_plt_x86_64_standard<32>(layout, got, got_plt, + got_irelative); +} + +template<> +Output_data_plt_x86_64<64>* +Target_x86_64<64>::do_make_data_plt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative) +{ + if (this->feature_1_ & elfcpp::GNU_PROPERTY_X86_FEATURE_1_IBT) + return new Output_data_plt_x86_64_ibt<64>(layout, got, got_plt, + got_irelative); + else if (parameters->options().bndplt()) + return new Output_data_plt_x86_64_bnd(layout, got, got_plt, + got_irelative); + else + return new Output_data_plt_x86_64_standard<64>(layout, got, got_plt, + got_irelative); +} + +template<> +Output_data_plt_x86_64<32>* +Target_x86_64<32>::do_make_data_plt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative, + unsigned int plt_count) +{ + if (this->feature_1_ & elfcpp::GNU_PROPERTY_X86_FEATURE_1_IBT) + return new Output_data_plt_x86_64_ibt<32>(layout, got, got_plt, + got_irelative, plt_count); + return new Output_data_plt_x86_64_standard<32>(layout, got, got_plt, + got_irelative, plt_count); +} + +template<> +Output_data_plt_x86_64<64>* +Target_x86_64<64>::do_make_data_plt(Layout* layout, + Output_data_got<64, false>* got, + Output_data_got_plt_x86_64* got_plt, + Output_data_space* got_irelative, + unsigned int plt_count) +{ + if (this->feature_1_ & elfcpp::GNU_PROPERTY_X86_FEATURE_1_IBT) + return new Output_data_plt_x86_64_ibt<64>(layout, got, got_plt, + got_irelative, plt_count); + else if (parameters->options().bndplt()) + return new Output_data_plt_x86_64_bnd(layout, got, got_plt, + got_irelative, plt_count); + else + return new Output_data_plt_x86_64_standard<64>(layout, got, got_plt, + got_irelative, + plt_count); +} + // Return the section for TLSDESC relocations. template @@ -1786,6 +3001,8 @@ template unsigned int Target_x86_64::first_plt_entry_offset() const { + if (this->plt_ == NULL) + return 0; return this->plt_->first_plt_entry_offset(); } @@ -1795,6 +3012,8 @@ template unsigned int Target_x86_64::plt_entry_size() const { + if (this->plt_ == NULL) + return 0; return this->plt_->get_plt_entry_size(); } @@ -2167,6 +3386,8 @@ Target_x86_64::Scan::get_reference_flags(unsigned int r_type) case elfcpp::R_X86_64_GOT32: case elfcpp::R_X86_64_GOTPCREL64: case elfcpp::R_X86_64_GOTPCREL: + case elfcpp::R_X86_64_GOTPCRELX: + case elfcpp::R_X86_64_REX_GOTPCRELX: case elfcpp::R_X86_64_GOTPLT64: // Absolute in GOT. return Symbol::ABSOLUTE_REF; @@ -2249,7 +3470,7 @@ Target_x86_64::Scan::check_non_pic(Relobj* object, unsigned int r_type, && !gsym->is_undefined() && !gsym->is_preemptible())) return; - /* Fall through. */ + // Fall through. case elfcpp::R_X86_64_32: // R_X86_64_32 is OK for x32. if (size == 32 && r_type == elfcpp::R_X86_64_32) @@ -2451,10 +3672,33 @@ Target_x86_64::Scan::local(Symbol_table* symtab, case elfcpp::R_X86_64_GOT32: case elfcpp::R_X86_64_GOTPCREL64: case elfcpp::R_X86_64_GOTPCREL: + case elfcpp::R_X86_64_GOTPCRELX: + case elfcpp::R_X86_64_REX_GOTPCRELX: case elfcpp::R_X86_64_GOTPLT64: { - // The symbol requires a GOT entry. + // The symbol requires a GOT section. Output_data_got<64, false>* got = target->got_section(symtab, layout); + + // If the relocation symbol isn't IFUNC, + // and is local, then we will convert + // mov foo@GOTPCREL(%rip), %reg + // to lea foo(%rip), %reg. + // in Relocate::relocate. + if (!parameters->incremental() + && (r_type == elfcpp::R_X86_64_GOTPCREL + || r_type == elfcpp::R_X86_64_GOTPCRELX + || r_type == elfcpp::R_X86_64_REX_GOTPCRELX) + && reloc.get_r_offset() >= 2 + && !is_ifunc) + { + section_size_type stype; + const unsigned char* view = object->section_contents(data_shndx, + &stype, true); + if (view[reloc.get_r_offset() - 2] == 0x8b) + break; + } + + // The symbol requires a GOT entry. unsigned int r_sym = elfcpp::elf_r_sym(reloc.get_r_info()); // For a STT_GNU_IFUNC symbol we want the PLT offset. That @@ -2657,7 +3901,11 @@ Target_x86_64::Scan::unsupported_reloc_global( // Returns true if this relocation type could be that of a function pointer. template inline bool -Target_x86_64::Scan::possible_function_pointer_reloc(unsigned int r_type) +Target_x86_64::Scan::possible_function_pointer_reloc( + Sized_relobj_file* src_obj, + unsigned int src_indx, + unsigned int r_offset, + unsigned int r_type) { switch (r_type) { @@ -2670,10 +3918,47 @@ Target_x86_64::Scan::possible_function_pointer_reloc(unsigned int r_type) case elfcpp::R_X86_64_GOT32: case elfcpp::R_X86_64_GOTPCREL64: case elfcpp::R_X86_64_GOTPCREL: + case elfcpp::R_X86_64_GOTPCRELX: + case elfcpp::R_X86_64_REX_GOTPCRELX: case elfcpp::R_X86_64_GOTPLT64: { return true; } + case elfcpp::R_X86_64_PC32: + { + // This relocation may be used both for function calls and + // for taking address of a function. We distinguish between + // them by checking the opcodes. + uint64_t sh_flags = src_obj->section_flags(src_indx); + bool is_executable = (sh_flags & elfcpp::SHF_EXECINSTR) != 0; + if (is_executable) + { + section_size_type stype; + const unsigned char* view = src_obj->section_contents(src_indx, + &stype, + true); + + // call + if (r_offset >= 1 + && view[r_offset - 1] == 0xe8) + return false; + + // jmp + if (r_offset >= 1 + && view[r_offset - 1] == 0xe9) + return false; + + // jo/jno/jb/jnb/je/jne/jna/ja/js/jns/jp/jnp/jl/jge/jle/jg + if (r_offset >= 2 + && view[r_offset - 2] == 0x0f + && view[r_offset - 1] >= 0x80 + && view[r_offset - 1] <= 0x8f) + return false; + } + + // Be conservative and treat all others as function pointers. + return true; + } } return false; } @@ -2688,18 +3973,21 @@ Target_x86_64::Scan::local_reloc_may_be_function_pointer( Symbol_table* , Layout* , Target_x86_64* , - Sized_relobj_file* , - unsigned int , + Sized_relobj_file* src_obj, + unsigned int src_indx, Output_section* , - const elfcpp::Rela& , + const elfcpp::Rela& reloc, unsigned int r_type, const elfcpp::Sym&) { // When building a shared library, do not fold any local symbols as it is // not possible to distinguish pointer taken versus a call by looking at // the relocation types. - return (parameters->options().shared() - || possible_function_pointer_reloc(r_type)); + if (parameters->options().shared()) + return true; + + return possible_function_pointer_reloc(src_obj, src_indx, + reloc.get_r_offset(), r_type); } // For safe ICF, scan a relocation for a global symbol to check if it @@ -2712,20 +4000,23 @@ Target_x86_64::Scan::global_reloc_may_be_function_pointer( Symbol_table*, Layout* , Target_x86_64* , - Sized_relobj_file* , - unsigned int , + Sized_relobj_file* src_obj, + unsigned int src_indx, Output_section* , - const elfcpp::Rela& , + const elfcpp::Rela& reloc, unsigned int r_type, Symbol* gsym) { // When building a shared library, do not fold symbols whose visibility // is hidden, internal or protected. - return ((parameters->options().shared() - && (gsym->visibility() == elfcpp::STV_INTERNAL - || gsym->visibility() == elfcpp::STV_PROTECTED - || gsym->visibility() == elfcpp::STV_HIDDEN)) - || possible_function_pointer_reloc(r_type)); + if (parameters->options().shared() + && (gsym->visibility() == elfcpp::STV_INTERNAL + || gsym->visibility() == elfcpp::STV_PROTECTED + || gsym->visibility() == elfcpp::STV_HIDDEN)) + return true; + + return possible_function_pointer_reloc(src_obj, src_indx, + reloc.get_r_offset(), r_type); } // Scan a relocation for a global symbol. @@ -2774,7 +4065,8 @@ Target_x86_64::Scan::global(Symbol_table* symtab, // Make a dynamic relocation if necessary. if (gsym->needs_dynamic_reloc(Scan::get_reference_flags(r_type))) { - if (gsym->may_need_copy_reloc()) + if (!parameters->options().output_is_position_independent() + && gsym->may_need_copy_reloc()) { target->copy_reloc(symtab, layout, object, data_shndx, output_section, gsym, reloc); @@ -2835,7 +4127,8 @@ Target_x86_64::Scan::global(Symbol_table* symtab, // Make a dynamic relocation if necessary. if (gsym->needs_dynamic_reloc(Scan::get_reference_flags(r_type))) { - if (gsym->may_need_copy_reloc()) + if (parameters->options().output_is_executable() + && gsym->may_need_copy_reloc()) { target->copy_reloc(symtab, layout, object, data_shndx, output_section, gsym, reloc); @@ -2856,10 +4149,41 @@ Target_x86_64::Scan::global(Symbol_table* symtab, case elfcpp::R_X86_64_GOT32: case elfcpp::R_X86_64_GOTPCREL64: case elfcpp::R_X86_64_GOTPCREL: + case elfcpp::R_X86_64_GOTPCRELX: + case elfcpp::R_X86_64_REX_GOTPCRELX: case elfcpp::R_X86_64_GOTPLT64: { // The symbol requires a GOT entry. Output_data_got<64, false>* got = target->got_section(symtab, layout); + + // If we convert this from + // mov foo@GOTPCREL(%rip), %reg + // to lea foo(%rip), %reg. + // OR + // if we convert + // (callq|jmpq) *foo@GOTPCRELX(%rip) to + // (callq|jmpq) foo + // in Relocate::relocate, then there is nothing to do here. + // We cannot make these optimizations in incremental linking mode, + // because we look at the opcode to decide whether or not to make + // change, and during an incremental update, the change may have + // already been applied. + + Lazy_view view(object, data_shndx); + size_t r_offset = reloc.get_r_offset(); + if (!parameters->incremental() + && r_offset >= 2 + && Target_x86_64::can_convert_mov_to_lea(gsym, r_type, + r_offset, &view)) + break; + + if (!parameters->incremental() + && r_offset >= 2 + && Target_x86_64::can_convert_callq_to_direct(gsym, r_type, + r_offset, + &view)) + break; + if (gsym->final_value_is_known()) { // For a STT_GNU_IFUNC symbol we want the PLT address. @@ -2921,11 +4245,6 @@ Target_x86_64::Scan::global(Symbol_table* symtab, } } } - // For GOTPLT64, we also need a PLT entry (but only if the - // symbol is not fully resolved). - if (r_type == elfcpp::R_X86_64_GOTPLT64 - && !gsym->final_value_is_known()) - target->make_plt_entry(symtab, layout, gsym); } break; @@ -2981,7 +4300,12 @@ Target_x86_64::Scan::global(Symbol_table* symtab, case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec case elfcpp::R_X86_64_TPOFF32: // Local-exec { - const bool is_final = gsym->final_value_is_known(); + // For the Initial-Exec model, we can treat undef symbols as final + // when building an executable. + const bool is_final = (gsym->final_value_is_known() || + (r_type == elfcpp::R_X86_64_GOTTPOFF && + gsym->is_undefined() && + parameters->options().output_is_executable())); const tls::Tls_optimization optimized_type = Target_x86_64::optimize_tls_reloc(is_final, r_type); switch (r_type) @@ -3110,15 +4434,16 @@ Target_x86_64::gc_process_relocs(Symbol_table* symtab, size_t local_symbol_count, const unsigned char* plocal_symbols) { + typedef gold::Default_classify_reloc + Classify_reloc; if (sh_type == elfcpp::SHT_REL) { return; } - gold::gc_process_relocs, elfcpp::SHT_RELA, - typename Target_x86_64::Scan, - typename Target_x86_64::Relocatable_size_for_reloc>( + gold::gc_process_relocs, Scan, + Classify_reloc>( symtab, layout, this, @@ -3148,6 +4473,9 @@ Target_x86_64::scan_relocs(Symbol_table* symtab, size_t local_symbol_count, const unsigned char* plocal_symbols) { + typedef gold::Default_classify_reloc + Classify_reloc; + if (sh_type == elfcpp::SHT_REL) { gold_error(_("%s: unsupported REL reloc section"), @@ -3155,8 +4483,7 @@ Target_x86_64::scan_relocs(Symbol_table* symtab, return; } - gold::scan_relocs, elfcpp::SHT_RELA, - typename Target_x86_64::Scan>( + gold::scan_relocs, Scan, Classify_reloc>( symtab, layout, this, @@ -3259,26 +4586,87 @@ Target_x86_64::do_finalize_sections( } } +// For x32, we need to handle PC-relative relocations using full 64-bit +// arithmetic, so that we can detect relocation overflows properly. +// This class overrides the pcrela32_check methods from the defaults in +// Relocate_functions in reloc.h. + +template +class X86_64_relocate_functions : public Relocate_functions +{ + public: + typedef Relocate_functions Base; + + // Do a simple PC relative relocation with the addend in the + // relocation. + static inline typename Base::Reloc_status + pcrela32_check(unsigned char* view, + typename elfcpp::Elf_types<64>::Elf_Addr value, + typename elfcpp::Elf_types<64>::Elf_Swxword addend, + typename elfcpp::Elf_types<64>::Elf_Addr address) + { + typedef typename elfcpp::Swap<32, false>::Valtype Valtype; + Valtype* wv = reinterpret_cast(view); + value = value + addend - address; + elfcpp::Swap<32, false>::writeval(wv, value); + return (Bits<32>::has_overflow(value) + ? Base::RELOC_OVERFLOW : Base::RELOC_OK); + } + + // Do a simple PC relative relocation with a Symbol_value with the + // addend in the relocation. + static inline typename Base::Reloc_status + pcrela32_check(unsigned char* view, + const Sized_relobj_file* object, + const Symbol_value* psymval, + typename elfcpp::Elf_types<64>::Elf_Swxword addend, + typename elfcpp::Elf_types<64>::Elf_Addr address) + { + typedef typename elfcpp::Swap<32, false>::Valtype Valtype; + Valtype* wv = reinterpret_cast(view); + typename elfcpp::Elf_types<64>::Elf_Addr value; + if (addend >= 0) + value = psymval->value(object, addend); + else + { + // For negative addends, get the symbol value without + // the addend, then add the addend using 64-bit arithmetic. + value = psymval->value(object, 0); + value += addend; + } + value -= address; + elfcpp::Swap<32, false>::writeval(wv, value); + return (Bits<32>::has_overflow(value) + ? Base::RELOC_OVERFLOW : Base::RELOC_OK); + } +}; + // Perform a relocation. template inline bool Target_x86_64::Relocate::relocate( const Relocate_info* relinfo, + unsigned int, Target_x86_64* target, Output_section*, size_t relnum, - const elfcpp::Rela& rela, - unsigned int r_type, + const unsigned char* preloc, const Sized_symbol* gsym, const Symbol_value* psymval, unsigned char* view, typename elfcpp::Elf_types::Elf_Addr address, section_size_type view_size) { + typedef X86_64_relocate_functions Reloc_funcs; + const elfcpp::Rela rela(preloc); + unsigned int r_type = elfcpp::elf_r_type(rela.get_r_info()); + if (this->skip_call_tls_get_addr_) { if ((r_type != elfcpp::R_X86_64_PLT32 + && r_type != elfcpp::R_X86_64_GOTPCREL + && r_type != elfcpp::R_X86_64_GOTPCRELX && r_type != elfcpp::R_X86_64_PLT32_BND && r_type != elfcpp::R_X86_64_PC32_BND && r_type != elfcpp::R_X86_64_PC32) @@ -3287,6 +4675,7 @@ Target_x86_64::Relocate::relocate( { gold_error_at_location(relinfo, relnum, rela.get_r_offset(), _("missing expected TLS relocation")); + this->skip_call_tls_get_addr_ = false; } else { @@ -3325,13 +4714,14 @@ Target_x86_64::Relocate::relocate( // We need to subtract the size of the GOT section to get // the actual offset to use in the relocation. bool have_got_offset = false; - unsigned int got_offset = 0; + // Since the actual offset is always negative, we use signed int to + // support 64-bit GOT relocations. + int got_offset = 0; switch (r_type) { case elfcpp::R_X86_64_GOT32: case elfcpp::R_X86_64_GOT64: case elfcpp::R_X86_64_GOTPLT64: - case elfcpp::R_X86_64_GOTPCREL: case elfcpp::R_X86_64_GOTPCREL64: if (gsym != NULL) { @@ -3352,6 +4742,8 @@ Target_x86_64::Relocate::relocate( break; } + typename Reloc_funcs::Reloc_status rstatus = Reloc_funcs::RELOC_OK; + switch (r_type) { case elfcpp::R_X86_64_NONE: @@ -3360,51 +4752,44 @@ Target_x86_64::Relocate::relocate( break; case elfcpp::R_X86_64_64: - Relocate_functions::rela64(view, object, psymval, addend); + Reloc_funcs::rela64(view, object, psymval, addend); break; case elfcpp::R_X86_64_PC64: - Relocate_functions::pcrela64(view, object, psymval, addend, + Reloc_funcs::pcrela64(view, object, psymval, addend, address); break; case elfcpp::R_X86_64_32: - // FIXME: we need to verify that value + addend fits into 32 bits: - // uint64_t x = value + addend; - // x == static_cast(static_cast(x)) - // Likewise for other <=32-bit relocations (but see R_X86_64_32S). - Relocate_functions::rela32(view, object, psymval, addend); + rstatus = Reloc_funcs::rela32_check(view, object, psymval, addend, + Reloc_funcs::CHECK_UNSIGNED); break; case elfcpp::R_X86_64_32S: - // FIXME: we need to verify that value + addend fits into 32 bits: - // int64_t x = value + addend; // note this quantity is signed! - // x == static_cast(static_cast(x)) - Relocate_functions::rela32(view, object, psymval, addend); + rstatus = Reloc_funcs::rela32_check(view, object, psymval, addend, + Reloc_funcs::CHECK_SIGNED); break; case elfcpp::R_X86_64_PC32: case elfcpp::R_X86_64_PC32_BND: - Relocate_functions::pcrela32(view, object, psymval, addend, - address); + rstatus = Reloc_funcs::pcrela32_check(view, object, psymval, addend, + address); break; case elfcpp::R_X86_64_16: - Relocate_functions::rela16(view, object, psymval, addend); + Reloc_funcs::rela16(view, object, psymval, addend); break; case elfcpp::R_X86_64_PC16: - Relocate_functions::pcrela16(view, object, psymval, addend, - address); + Reloc_funcs::pcrela16(view, object, psymval, addend, address); break; case elfcpp::R_X86_64_8: - Relocate_functions::rela8(view, object, psymval, addend); + Reloc_funcs::rela8(view, object, psymval, addend); break; case elfcpp::R_X86_64_PC8: - Relocate_functions::pcrela8(view, object, psymval, addend, - address); + Reloc_funcs::pcrela8(view, object, psymval, addend, address); break; case elfcpp::R_X86_64_PLT32: @@ -3418,8 +4803,8 @@ Target_x86_64::Relocate::relocate( // Note: while this code looks the same as for R_X86_64_PC32, it // behaves differently because psymval was set to point to // the PLT entry, rather than the symbol, in Scan::global(). - Relocate_functions::pcrela32(view, object, psymval, addend, - address); + rstatus = Reloc_funcs::pcrela32_check(view, object, psymval, addend, + address); break; case elfcpp::R_X86_64_PLTOFF64: @@ -3428,15 +4813,15 @@ Target_x86_64::Relocate::relocate( gold_assert(gsym->has_plt_offset() || gsym->final_value_is_known()); typename elfcpp::Elf_types::Elf_Addr got_address; - got_address = target->got_section(NULL, NULL)->address(); - Relocate_functions::rela64(view, object, psymval, - addend - got_address); + // This is the address of GLOBAL_OFFSET_TABLE. + got_address = target->got_plt_section()->address(); + Reloc_funcs::rela64(view, object, psymval, addend - got_address); } break; case elfcpp::R_X86_64_GOT32: gold_assert(have_got_offset); - Relocate_functions::rela32(view, got_offset, addend); + Reloc_funcs::rela32(view, got_offset, addend); break; case elfcpp::R_X86_64_GOTPC32: @@ -3444,16 +4829,16 @@ Target_x86_64::Relocate::relocate( gold_assert(gsym); typename elfcpp::Elf_types::Elf_Addr value; value = target->got_plt_section()->address(); - Relocate_functions::pcrela32(view, value, addend, address); + Reloc_funcs::pcrela32_check(view, value, addend, address); } break; case elfcpp::R_X86_64_GOT64: - // The ABI doc says "Like GOT64, but indicates a PLT entry is needed." - // Since we always add a PLT entry, this is equivalent. case elfcpp::R_X86_64_GOTPLT64: + // R_X86_64_GOTPLT64 is obsolete and treated the same as + // GOT64. gold_assert(have_got_offset); - Relocate_functions::rela64(view, got_offset, addend); + Reloc_funcs::rela64(view, got_offset, addend); break; case elfcpp::R_X86_64_GOTPC64: @@ -3461,25 +4846,98 @@ Target_x86_64::Relocate::relocate( gold_assert(gsym); typename elfcpp::Elf_types::Elf_Addr value; value = target->got_plt_section()->address(); - Relocate_functions::pcrela64(view, value, addend, address); + Reloc_funcs::pcrela64(view, value, addend, address); } break; case elfcpp::R_X86_64_GOTOFF64: { - typename elfcpp::Elf_types::Elf_Addr value; - value = (psymval->value(object, 0) - - target->got_plt_section()->address()); - Relocate_functions::rela64(view, value, addend); + typename elfcpp::Elf_types::Elf_Addr reladdr; + reladdr = target->got_plt_section()->address(); + Reloc_funcs::pcrela64(view, object, psymval, addend, reladdr); } break; case elfcpp::R_X86_64_GOTPCREL: + case elfcpp::R_X86_64_GOTPCRELX: + case elfcpp::R_X86_64_REX_GOTPCRELX: { - gold_assert(have_got_offset); - typename elfcpp::Elf_types::Elf_Addr value; - value = target->got_plt_section()->address() + got_offset; - Relocate_functions::pcrela32(view, value, addend, address); + // Convert + // mov foo@GOTPCREL(%rip), %reg + // to lea foo(%rip), %reg. + // if possible. + if (!parameters->incremental() + && ((gsym == NULL + && rela.get_r_offset() >= 2 + && view[-2] == 0x8b + && !psymval->is_ifunc_symbol()) + || (gsym != NULL + && rela.get_r_offset() >= 2 + && Target_x86_64::can_convert_mov_to_lea(gsym, r_type, + 0, &view)))) + { + view[-2] = 0x8d; + Reloc_funcs::pcrela32(view, object, psymval, addend, address); + } + // Convert + // callq *foo@GOTPCRELX(%rip) to + // addr32 callq foo + // and jmpq *foo@GOTPCRELX(%rip) to + // jmpq foo + // nop + else if (!parameters->incremental() + && gsym != NULL + && rela.get_r_offset() >= 2 + && Target_x86_64::can_convert_callq_to_direct(gsym, + r_type, + 0, &view)) + { + if (view[-1] == 0x15) + { + // Convert callq *foo@GOTPCRELX(%rip) to addr32 callq. + // Opcode of addr32 is 0x67 and opcode of direct callq is 0xe8. + view[-2] = 0x67; + view[-1] = 0xe8; + // Convert GOTPCRELX to 32-bit pc relative reloc. + Reloc_funcs::pcrela32(view, object, psymval, addend, address); + } + else + { + // Convert jmpq *foo@GOTPCRELX(%rip) to + // jmpq foo + // nop + // The opcode of direct jmpq is 0xe9. + view[-2] = 0xe9; + // The opcode of nop is 0x90. + view[3] = 0x90; + // Convert GOTPCRELX to 32-bit pc relative reloc. jmpq is rip + // relative and since the instruction following the jmpq is now + // the nop, offset the address by 1 byte. The start of the + // relocation also moves ahead by 1 byte. + Reloc_funcs::pcrela32(&view[-1], object, psymval, addend, + address - 1); + } + } + else + { + if (gsym != NULL) + { + gold_assert(gsym->has_got_offset(GOT_TYPE_STANDARD)); + got_offset = (gsym->got_offset(GOT_TYPE_STANDARD) + - target->got_size()); + } + else + { + unsigned int r_sym = elfcpp::elf_r_sym(rela.get_r_info()); + gold_assert(object->local_has_got_offset(r_sym, + GOT_TYPE_STANDARD)); + got_offset = (object->local_got_offset(r_sym, GOT_TYPE_STANDARD) + - target->got_size()); + } + typename elfcpp::Elf_types::Elf_Addr value; + value = target->got_plt_section()->address() + got_offset; + Reloc_funcs::pcrela32_check(view, value, addend, address); + } } break; @@ -3488,7 +4946,7 @@ Target_x86_64::Relocate::relocate( gold_assert(have_got_offset); typename elfcpp::Elf_types::Elf_Addr value; value = target->got_plt_section()->address() + got_offset; - Relocate_functions::pcrela64(view, value, addend, address); + Reloc_funcs::pcrela64(view, value, addend, address); } break; @@ -3528,6 +4986,32 @@ Target_x86_64::Relocate::relocate( break; } + if (rstatus == Reloc_funcs::RELOC_OVERFLOW) + { + if (gsym == NULL) + { + unsigned int r_sym = elfcpp::elf_r_sym(rela.get_r_info()); + gold_error_at_location(relinfo, relnum, rela.get_r_offset(), + _("relocation overflow: " + "reference to local symbol %u in %s"), + r_sym, object->name().c_str()); + } + else if (gsym->is_defined() && gsym->source() == Symbol::FROM_OBJECT) + { + gold_error_at_location(relinfo, relnum, rela.get_r_offset(), + _("relocation overflow: " + "reference to '%s' defined in %s"), + gsym->name(), + gsym->object()->name().c_str()); + } + else + { + gold_error_at_location(relinfo, relnum, rela.get_r_offset(), + _("relocation overflow: reference to '%s'"), + gsym->name()); + } + } + return true; } @@ -3608,7 +5092,7 @@ Target_x86_64::Relocate::relocate_tls( if (optimized_type == tls::TLSOPT_TO_IE) { value = target->got_plt_section()->address() + got_offset; - this->tls_gd_to_ie(relinfo, relnum, tls_segment, rela, r_type, + this->tls_gd_to_ie(relinfo, relnum, rela, r_type, value, view, address, view_size); break; } @@ -3675,14 +5159,8 @@ Target_x86_64::Relocate::relocate_tls( } if (optimized_type == tls::TLSOPT_TO_IE) { - if (tls_segment == NULL) - { - gold_assert(parameters->errors()->error_count() > 0 - || issue_undefined_symbol_error(gsym)); - return; - } value = target->got_plt_section()->address() + got_offset; - this->tls_desc_gd_to_ie(relinfo, relnum, tls_segment, + this->tls_desc_gd_to_ie(relinfo, relnum, rela, r_type, value, view, address, view_size); break; @@ -3774,7 +5252,17 @@ Target_x86_64::Relocate::relocate_tls( break; case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec - if (optimized_type == tls::TLSOPT_TO_LE) + if (gsym != NULL + && gsym->is_undefined() + && parameters->options().output_is_executable()) + { + Target_x86_64::Relocate::tls_ie_to_le(relinfo, relnum, + NULL, rela, + r_type, value, view, + view_size); + break; + } + else if (optimized_type == tls::TLSOPT_TO_LE) { if (tls_segment == NULL) { @@ -3838,7 +5326,6 @@ inline void Target_x86_64::Relocate::tls_gd_to_ie( const Relocate_info* relinfo, size_t relnum, - Output_segment*, const elfcpp::Rela& rela, unsigned int, typename elfcpp::Elf_types::Elf_Addr value, @@ -3848,16 +5335,23 @@ Target_x86_64::Relocate::tls_gd_to_ie( { // For SIZE == 64: // .byte 0x66; leaq foo@tlsgd(%rip),%rdi; - // .word 0x6666; rex64; call __tls_get_addr + // .word 0x6666; rex64; call __tls_get_addr@PLT + // ==> movq %fs:0,%rax; addq x@gottpoff(%rip),%rax + // .byte 0x66; leaq foo@tlsgd(%rip),%rdi; + // .word 0x66; rex64; call *__tls_get_addr@GOTPCREL(%rip) // ==> movq %fs:0,%rax; addq x@gottpoff(%rip),%rax // For SIZE == 32: // leaq foo@tlsgd(%rip),%rdi; - // .word 0x6666; rex64; call __tls_get_addr + // .word 0x6666; rex64; call __tls_get_addr@PLT + // ==> movl %fs:0,%eax; addq x@gottpoff(%rip),%rax + // leaq foo@tlsgd(%rip),%rdi; + // .word 0x66; rex64; call *__tls_get_addr@GOTPCREL(%rip) // ==> movl %fs:0,%eax; addq x@gottpoff(%rip),%rax tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 12); tls::check_tls(relinfo, relnum, rela.get_r_offset(), - (memcmp(view + 4, "\x66\x66\x48\xe8", 4) == 0)); + (memcmp(view + 4, "\x66\x66\x48\xe8", 4) == 0 + || memcmp(view + 4, "\x66\x48\xff", 3) == 0)); if (size == 64) { @@ -3904,16 +5398,23 @@ Target_x86_64::Relocate::tls_gd_to_le( { // For SIZE == 64: // .byte 0x66; leaq foo@tlsgd(%rip),%rdi; - // .word 0x6666; rex64; call __tls_get_addr + // .word 0x6666; rex64; call __tls_get_addr@PLT + // ==> movq %fs:0,%rax; leaq x@tpoff(%rax),%rax + // .byte 0x66; leaq foo@tlsgd(%rip),%rdi; + // .word 0x66; rex64; call *__tls_get_addr@GOTPCREL(%rip) // ==> movq %fs:0,%rax; leaq x@tpoff(%rax),%rax // For SIZE == 32: // leaq foo@tlsgd(%rip),%rdi; - // .word 0x6666; rex64; call __tls_get_addr + // .word 0x6666; rex64; call __tls_get_addr@PLT + // ==> movl %fs:0,%eax; leaq x@tpoff(%rax),%rax + // leaq foo@tlsgd(%rip),%rdi; + // .word 0x66; rex64; call *__tls_get_addr@GOTPCREL(%rip) // ==> movl %fs:0,%eax; leaq x@tpoff(%rax),%rax tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 12); tls::check_tls(relinfo, relnum, rela.get_r_offset(), - (memcmp(view + 4, "\x66\x66\x48\xe8", 4) == 0)); + (memcmp(view + 4, "\x66\x66\x48\xe8", 4) == 0 + || memcmp(view + 4, "\x66\x48\xff", 3) == 0)); if (size == 64) { @@ -3950,7 +5451,6 @@ inline void Target_x86_64::Relocate::tls_desc_gd_to_ie( const Relocate_info* relinfo, size_t relnum, - Output_segment*, const elfcpp::Rela& rela, unsigned int r_type, typename elfcpp::Elf_types::Elf_Addr value, @@ -4042,6 +5542,13 @@ Target_x86_64::Relocate::tls_ld_to_le( // For SIZE == 32: // ... leq foo@dtpoff(%rax),%reg // ==> nopl 0x0(%rax); movl %fs:0,%eax ... leaq x@tpoff(%rax),%rdx + // leaq foo@tlsld(%rip),%rdi; call *__tls_get_addr@GOTPCREL(%rip) + // For SIZE == 64: + // ... leq foo@dtpoff(%rax),%reg + // ==> .word 0x6666; .byte 0x6666; movq %fs:0,%rax ... leaq x@tpoff(%rax),%rdx + // For SIZE == 32: + // ... leq foo@dtpoff(%rax),%reg + // ==> nopw 0x0(%rax); movl %fs:0,%eax ... leaq x@tpoff(%rax),%rdx tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, -3); tls::check_range(relinfo, relnum, rela.get_r_offset(), view_size, 9); @@ -4049,12 +5556,25 @@ Target_x86_64::Relocate::tls_ld_to_le( tls::check_tls(relinfo, relnum, rela.get_r_offset(), view[-3] == 0x48 && view[-2] == 0x8d && view[-1] == 0x3d); - tls::check_tls(relinfo, relnum, rela.get_r_offset(), view[4] == 0xe8); + tls::check_tls(relinfo, relnum, rela.get_r_offset(), + view[4] == 0xe8 || view[4] == 0xff); - if (size == 64) - memcpy(view - 3, "\x66\x66\x66\x64\x48\x8b\x04\x25\0\0\0\0", 12); + if (view[4] == 0xe8) + { + if (size == 64) + memcpy(view - 3, "\x66\x66\x66\x64\x48\x8b\x04\x25\0\0\0\0", 12); + else + memcpy(view - 3, "\x0f\x1f\x40\x00\x64\x8b\x04\x25\0\0\0\0", 12); + } else - memcpy(view - 3, "\x0f\x1f\x40\x00\x64\x8b\x04\x25\0\0\0\0", 12); + { + if (size == 64) + memcpy(view - 3, "\x66\x66\x66\x66\x64\x48\x8b\x04\x25\0\0\0\0", + 13); + else + memcpy(view - 3, "\x66\x0f\x1f\x40\x00\x64\x8b\x04\x25\0\0\0\0", + 13); + } // The next reloc should be a PLT32 reloc against __tls_get_addr. // We can skip it. @@ -4095,6 +5615,8 @@ Target_x86_64::Relocate::tls_ie_to_le( // movq if (op1 == 0x4c) view[-3] = 0x49; + else if (size == 32 && op1 == 0x44) + view[-3] = 0x41; view[-2] = 0xc7; view[-1] = 0xc0 | reg; } @@ -4103,6 +5625,8 @@ Target_x86_64::Relocate::tls_ie_to_le( // Special handling for %rsp. if (op1 == 0x4c) view[-3] = 0x49; + else if (size == 32 && op1 == 0x44) + view[-3] = 0x41; view[-2] = 0x81; view[-1] = 0xc0 | reg; } @@ -4111,11 +5635,14 @@ Target_x86_64::Relocate::tls_ie_to_le( // addq if (op1 == 0x4c) view[-3] = 0x4d; + else if (size == 32 && op1 == 0x44) + view[-3] = 0x45; view[-2] = 0x8d; view[-1] = 0x80 | reg | (reg << 3); } - value -= tls_segment->memsz(); + if (tls_segment != NULL) + value -= tls_segment->memsz(); Relocate_functions::rela32(view, value, 0); } @@ -4135,11 +5662,13 @@ Target_x86_64::relocate_section( section_size_type view_size, const Reloc_symbol_changes* reloc_symbol_changes) { + typedef gold::Default_classify_reloc + Classify_reloc; + gold_assert(sh_type == elfcpp::SHT_RELA); - gold::relocate_section, elfcpp::SHT_RELA, - typename Target_x86_64::Relocate, - gold::Default_comdat_behavior>( + gold::relocate_section, Relocate, + gold::Default_comdat_behavior, Classify_reloc>( relinfo, this, prelocs, @@ -4180,84 +5709,50 @@ Target_x86_64::apply_relocation( view_size); } -// Return the size of a relocation while scanning during a relocatable -// link. +// Scan the relocs during a relocatable link. template -unsigned int -Target_x86_64::Relocatable_size_for_reloc::get_size_for_reloc( - unsigned int r_type, - Relobj* object) +void +Target_x86_64::scan_relocatable_relocs( + Symbol_table* symtab, + Layout* layout, + Sized_relobj_file* object, + unsigned int data_shndx, + unsigned int sh_type, + const unsigned char* prelocs, + size_t reloc_count, + Output_section* output_section, + bool needs_special_offset_handling, + size_t local_symbol_count, + const unsigned char* plocal_symbols, + Relocatable_relocs* rr) { - switch (r_type) - { - case elfcpp::R_X86_64_NONE: - case elfcpp::R_X86_64_GNU_VTINHERIT: - case elfcpp::R_X86_64_GNU_VTENTRY: - case elfcpp::R_X86_64_TLSGD: // Global-dynamic - case elfcpp::R_X86_64_GOTPC32_TLSDESC: // Global-dynamic (from ~oliva url) - case elfcpp::R_X86_64_TLSDESC_CALL: - case elfcpp::R_X86_64_TLSLD: // Local-dynamic - case elfcpp::R_X86_64_DTPOFF32: - case elfcpp::R_X86_64_DTPOFF64: - case elfcpp::R_X86_64_GOTTPOFF: // Initial-exec - case elfcpp::R_X86_64_TPOFF32: // Local-exec - return 0; - - case elfcpp::R_X86_64_64: - case elfcpp::R_X86_64_PC64: - case elfcpp::R_X86_64_GOTOFF64: - case elfcpp::R_X86_64_GOTPC64: - case elfcpp::R_X86_64_PLTOFF64: - case elfcpp::R_X86_64_GOT64: - case elfcpp::R_X86_64_GOTPCREL64: - case elfcpp::R_X86_64_GOTPCREL: - case elfcpp::R_X86_64_GOTPLT64: - return 8; - - case elfcpp::R_X86_64_32: - case elfcpp::R_X86_64_32S: - case elfcpp::R_X86_64_PC32: - case elfcpp::R_X86_64_PC32_BND: - case elfcpp::R_X86_64_PLT32: - case elfcpp::R_X86_64_PLT32_BND: - case elfcpp::R_X86_64_GOTPC32: - case elfcpp::R_X86_64_GOT32: - return 4; - - case elfcpp::R_X86_64_16: - case elfcpp::R_X86_64_PC16: - return 2; + typedef gold::Default_classify_reloc + Classify_reloc; + typedef gold::Default_scan_relocatable_relocs + Scan_relocatable_relocs; - case elfcpp::R_X86_64_8: - case elfcpp::R_X86_64_PC8: - return 1; - - case elfcpp::R_X86_64_COPY: - case elfcpp::R_X86_64_GLOB_DAT: - case elfcpp::R_X86_64_JUMP_SLOT: - case elfcpp::R_X86_64_RELATIVE: - case elfcpp::R_X86_64_IRELATIVE: - // These are outstanding tls relocs, which are unexpected when linking - case elfcpp::R_X86_64_TPOFF64: - case elfcpp::R_X86_64_DTPMOD64: - case elfcpp::R_X86_64_TLSDESC: - object->error(_("unexpected reloc %u in object file"), r_type); - return 0; + gold_assert(sh_type == elfcpp::SHT_RELA); - case elfcpp::R_X86_64_SIZE32: - case elfcpp::R_X86_64_SIZE64: - default: - object->error(_("unsupported reloc %u against local symbol"), r_type); - return 0; - } + gold::scan_relocatable_relocs( + symtab, + layout, + object, + data_shndx, + prelocs, + reloc_count, + output_section, + needs_special_offset_handling, + local_symbol_count, + plocal_symbols, + rr); } -// Scan the relocs during a relocatable link. +// Scan the relocs for --emit-relocs. template void -Target_x86_64::scan_relocatable_relocs( +Target_x86_64::emit_relocs_scan( Symbol_table* symtab, Layout* layout, Sized_relobj_file* object, @@ -4268,16 +5763,17 @@ Target_x86_64::scan_relocatable_relocs( Output_section* output_section, bool needs_special_offset_handling, size_t local_symbol_count, - const unsigned char* plocal_symbols, + const unsigned char* plocal_syms, Relocatable_relocs* rr) { - gold_assert(sh_type == elfcpp::SHT_RELA); + typedef gold::Default_classify_reloc + Classify_reloc; + typedef gold::Default_emit_relocs_strategy + Emit_relocs_strategy; - typedef gold::Default_scan_relocatable_relocs Scan_relocatable_relocs; + gold_assert(sh_type == elfcpp::SHT_RELA); - gold::scan_relocatable_relocs( + gold::scan_relocatable_relocs( symtab, layout, object, @@ -4287,7 +5783,7 @@ Target_x86_64::scan_relocatable_relocs( output_section, needs_special_offset_handling, local_symbol_count, - plocal_symbols, + plocal_syms, rr); } @@ -4302,22 +5798,23 @@ Target_x86_64::relocate_relocs( size_t reloc_count, Output_section* output_section, typename elfcpp::Elf_types::Elf_Off offset_in_output_section, - const Relocatable_relocs* rr, unsigned char* view, typename elfcpp::Elf_types::Elf_Addr view_address, section_size_type view_size, unsigned char* reloc_view, section_size_type reloc_view_size) { + typedef gold::Default_classify_reloc + Classify_reloc; + gold_assert(sh_type == elfcpp::SHT_RELA); - gold::relocate_relocs( + gold::relocate_relocs( relinfo, prelocs, reloc_count, output_section, offset_in_output_section, - rr, view, view_address, view_size, @@ -4452,35 +5949,60 @@ Target_x86_64::do_ehframe_datarel_base() const // code. We have to change the function so that it always ensures // that it has enough stack space to run some random function. +static const unsigned char cmp_insn_32[] = { 0x64, 0x3b, 0x24, 0x25 }; +static const unsigned char lea_r10_insn_32[] = { 0x44, 0x8d, 0x94, 0x24 }; +static const unsigned char lea_r11_insn_32[] = { 0x44, 0x8d, 0x9c, 0x24 }; + +static const unsigned char cmp_insn_64[] = { 0x64, 0x48, 0x3b, 0x24, 0x25 }; +static const unsigned char lea_r10_insn_64[] = { 0x4c, 0x8d, 0x94, 0x24 }; +static const unsigned char lea_r11_insn_64[] = { 0x4c, 0x8d, 0x9c, 0x24 }; + template void Target_x86_64::do_calls_non_split(Relobj* object, unsigned int shndx, section_offset_type fnoffset, section_size_type fnsize, + const unsigned char*, + size_t, unsigned char* view, section_size_type view_size, std::string* from, std::string* to) const { + const char* const cmp_insn = reinterpret_cast + (size == 32 ? cmp_insn_32 : cmp_insn_64); + const char* const lea_r10_insn = reinterpret_cast + (size == 32 ? lea_r10_insn_32 : lea_r10_insn_64); + const char* const lea_r11_insn = reinterpret_cast + (size == 32 ? lea_r11_insn_32 : lea_r11_insn_64); + + const size_t cmp_insn_len = + (size == 32 ? sizeof(cmp_insn_32) : sizeof(cmp_insn_64)); + const size_t lea_r10_insn_len = + (size == 32 ? sizeof(lea_r10_insn_32) : sizeof(lea_r10_insn_64)); + const size_t lea_r11_insn_len = + (size == 32 ? sizeof(lea_r11_insn_32) : sizeof(lea_r11_insn_64)); + const size_t nop_len = (size == 32 ? 7 : 8); + // The function starts with a comparison of the stack pointer and a // field in the TCB. This is followed by a jump. // cmp %fs:NN,%rsp - if (this->match_view(view, view_size, fnoffset, "\x64\x48\x3b\x24\x25", 5) - && fnsize > 9) + if (this->match_view(view, view_size, fnoffset, cmp_insn, cmp_insn_len) + && fnsize > nop_len + 1) { // We will call __morestack if the carry flag is set after this // comparison. We turn the comparison into an stc instruction // and some nops. view[fnoffset] = '\xf9'; - this->set_view_to_nop(view, view_size, fnoffset + 1, 8); + this->set_view_to_nop(view, view_size, fnoffset + 1, nop_len); } // lea NN(%rsp),%r10 // lea NN(%rsp),%r11 else if ((this->match_view(view, view_size, fnoffset, - "\x4c\x8d\x94\x24", 4) + lea_r10_insn, lea_r10_insn_len) || this->match_view(view, view_size, fnoffset, - "\x4c\x8d\x9c\x24", 4)) + lea_r11_insn, lea_r11_insn_len)) && fnsize > 8) { // This is loading an offset from the stack pointer for a @@ -4671,7 +6193,9 @@ const Target::Target_info Target_x86_64_nacl<64>::x86_64_nacl_info = elfcpp::SHF_X86_64_LARGE, // large_common_section_flags NULL, // attributes_section NULL, // attributes_vendor - "_start" // entry_symbol_name + "_start", // entry_symbol_name + 32, // hash_entry_size + elfcpp::SHT_X86_64_UNWIND, // unwind_section_type }; template<> @@ -4698,7 +6222,9 @@ const Target::Target_info Target_x86_64_nacl<32>::x86_64_nacl_info = elfcpp::SHF_X86_64_LARGE, // large_common_section_flags NULL, // attributes_section NULL, // attributes_vendor - "_start" // entry_symbol_name + "_start", // entry_symbol_name + 32, // hash_entry_size + elfcpp::SHT_X86_64_UNWIND, // unwind_section_type }; #define NACLMASK 0xe0 // 32-byte alignment mask.