From: Alan Modra Date: Tue, 17 Jun 2003 13:43:02 +0000 (+0000) Subject: * elf64-ppc.c (struct ppc_link_hash_table): Reinstate top_index. X-Git-Url: https://git.libre-soc.org/?a=commitdiff_plain;h=734b6cf9ae66bb2f7ced350c570aa339fb5ab707;p=binutils-gdb.git * elf64-ppc.c (struct ppc_link_hash_table): Reinstate top_index. Restore previous input_list type. (ppc64_elf_link_hash_table_create): Undo last change. (ppc64_elf_setup_section_lists): Reinstate code setting up input lists per output section, but don't bother with bfd_abs_section marker. (ppc64_elf_next_input_section): Adjust for multiple input section lists. (group_sections): Likewise. --- diff --git a/bfd/ChangeLog b/bfd/ChangeLog index 46ac7182014..288c815e85a 100644 --- a/bfd/ChangeLog +++ b/bfd/ChangeLog @@ -1,3 +1,14 @@ +2003-06-17 Alan Modra + + * elf64-ppc.c (struct ppc_link_hash_table): Reinstate top_index. + Restore previous input_list type. + (ppc64_elf_link_hash_table_create): Undo last change. + (ppc64_elf_setup_section_lists): Reinstate code setting up input lists + per output section, but don't bother with bfd_abs_section marker. + (ppc64_elf_next_input_section): Adjust for multiple input section + lists. + (group_sections): Likewise. + 2003-06-17 Jakub Jelinek * elf64-ppc.c (ppc64_elf_relocation_section): Ensure diff --git a/bfd/elf64-ppc.c b/bfd/elf64-ppc.c index e7c0836183a..97fd482a7d1 100644 --- a/bfd/elf64-ppc.c +++ b/bfd/elf64-ppc.c @@ -2773,8 +2773,11 @@ struct ppc_link_hash_table /* Temp used when calculating TOC pointers. */ bfd_vma toc_curr; - /* List of input code sections used by ppc64_elf_size_stubs. */ - asection *input_list; + /* Highest output section index. */ + int top_index; + + /* List of input sections for each output section. */ + asection **input_list; /* Short-cuts to get to dynamic linker sections. */ asection *sgot; @@ -3053,7 +3056,6 @@ ppc64_elf_link_hash_table_create (abfd) htab->no_multi_toc = 0; htab->multi_toc_needed = 0; htab->toc_curr = 0; - htab->input_list = NULL; htab->sgot = NULL; htab->srelgot = NULL; htab->splt = NULL; @@ -6500,8 +6502,9 @@ ppc64_elf_setup_section_lists (output_bfd, info) struct bfd_link_info *info; { bfd *input_bfd; - int top_id, id; + int top_id, top_index, id; asection *section; + asection **input_list; bfd_size_type amt; struct ppc_link_hash_table *htab = ppc_hash_table (info); @@ -6532,6 +6535,25 @@ ppc64_elf_setup_section_lists (output_bfd, info) htab->stub_group[id].toc_off = TOC_BASE_OFF; elf_gp (output_bfd) = htab->toc_curr = ppc64_elf_toc (output_bfd); + + /* We can't use output_bfd->section_count here to find the top output + section index as some sections may have been removed, and + _bfd_strip_section_from_output doesn't renumber the indices. */ + for (section = output_bfd->sections, top_index = 0; + section != NULL; + section = section->next) + { + if (top_index < section->index) + top_index = section->index; + } + + htab->top_index = top_index; + amt = sizeof (asection *) * (top_index + 1); + input_list = (asection **) bfd_zmalloc (amt); + htab->input_list = input_list; + if (input_list == NULL) + return -1; + return 1; } @@ -6588,14 +6610,16 @@ ppc64_elf_next_input_section (info, isec) { struct ppc_link_hash_table *htab = ppc_hash_table (info); - if ((isec->output_section->flags & SEC_CODE) != 0) + if ((isec->output_section->flags & SEC_CODE) != 0 + && isec->output_section->index <= htab->top_index) { + asection **list = htab->input_list + isec->output_section->index; /* Steal the link_sec pointer for our list. */ #define PREV_SEC(sec) (htab->stub_group[(sec)->id].link_sec) /* This happens to make the list in reverse order, which is what we want. */ - PREV_SEC (isec) = htab->input_list; - htab->input_list = isec; + PREV_SEC (isec) = *list; + *list = isec; } /* If a code section has a function that uses the TOC then we need @@ -6624,73 +6648,73 @@ group_sections (htab, stub_group_size, stubs_always_before_branch) bfd_size_type stub_group_size; bfd_boolean stubs_always_before_branch; { - asection *tail = htab->input_list; - while (tail != NULL) + asection **list = htab->input_list + htab->top_index; + do { - asection *curr; - asection *prev; - bfd_size_type total; - bfd_boolean big_sec; - bfd_vma curr_toc; - - curr = tail; - if (tail->_cooked_size) - total = tail->_cooked_size; - else - total = tail->_raw_size; - big_sec = total >= stub_group_size; - curr_toc = htab->stub_group[tail->id].toc_off; - - while ((prev = PREV_SEC (curr)) != NULL - && ((total += (curr->output_section->vma - + curr->output_offset - - prev->output_section->vma - - prev->output_offset)) - < stub_group_size) - && htab->stub_group[prev->id].toc_off == curr_toc) - curr = prev; - - /* OK, the size from the start of CURR to the end is less - than stub_group_size and thus can be handled by one stub - section. (or the tail section is itself larger than - stub_group_size, in which case we may be toast.) We - should really be keeping track of the total size of stubs - added here, as stubs contribute to the final output - section size. That's a little tricky, and this way will - only break if stubs added make the total size more than - 2^25, ie. for the default stub_group_size, if stubs total - more than 2097152 bytes, or nearly 75000 plt call stubs. */ - do + asection *tail = *list; + while (tail != NULL) { - prev = PREV_SEC (tail); - /* Set up this stub group. */ - htab->stub_group[tail->id].link_sec = curr; - } - while (tail != curr && (tail = prev) != NULL); - - /* But wait, there's more! Input sections up to stub_group_size - bytes before the stub section can be handled by it too. - Don't do this if we have a really large section after the - stubs, as adding more stubs increases the chance that - branches may not reach into the stub section. */ - if (!stubs_always_before_branch && !big_sec) - { - total = 0; - while (prev != NULL - && ((total += (tail->output_section->vma - + tail->output_offset - - prev->output_section->vma - - prev->output_offset)) + asection *curr; + asection *prev; + bfd_size_type total; + bfd_boolean big_sec; + bfd_vma curr_toc; + + curr = tail; + if (tail->_cooked_size) + total = tail->_cooked_size; + else + total = tail->_raw_size; + big_sec = total >= stub_group_size; + curr_toc = htab->stub_group[tail->id].toc_off; + + while ((prev = PREV_SEC (curr)) != NULL + && ((total += curr->output_offset - prev->output_offset) < stub_group_size) && htab->stub_group[prev->id].toc_off == curr_toc) + curr = prev; + + /* OK, the size from the start of CURR to the end is less + than stub_group_size and thus can be handled by one stub + section. (or the tail section is itself larger than + stub_group_size, in which case we may be toast.) We + should really be keeping track of the total size of stubs + added here, as stubs contribute to the final output + section size. That's a little tricky, and this way will + only break if stubs added make the total size more than + 2^25, ie. for the default stub_group_size, if stubs total + more than 2097152 bytes, or nearly 75000 plt call stubs. */ + do { - tail = prev; prev = PREV_SEC (tail); + /* Set up this stub group. */ htab->stub_group[tail->id].link_sec = curr; } + while (tail != curr && (tail = prev) != NULL); + + /* But wait, there's more! Input sections up to stub_group_size + bytes before the stub section can be handled by it too. + Don't do this if we have a really large section after the + stubs, as adding more stubs increases the chance that + branches may not reach into the stub section. */ + if (!stubs_always_before_branch && !big_sec) + { + total = 0; + while (prev != NULL + && ((total += tail->output_offset - prev->output_offset) + < stub_group_size) + && htab->stub_group[prev->id].toc_off == curr_toc) + { + tail = prev; + prev = PREV_SEC (tail); + htab->stub_group[tail->id].link_sec = curr; + } + } + tail = prev; } - tail = prev; } + while (list-- != htab->input_list); + free (htab->input_list); #undef PREV_SEC }