bfd/
[deliverable/binutils-gdb.git] / bfd / elf32-spu.c
1 /* SPU specific support for 32-bit ELF
2
3 Copyright 2006, 2007, 2008, 2009, 2010, 2011, 2012
4 Free Software Foundation, Inc.
5
6 This file is part of BFD, the Binary File Descriptor library.
7
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3 of the License, or
11 (at your option) any later version.
12
13 This program is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
17
18 You should have received a copy of the GNU General Public License along
19 with this program; if not, write to the Free Software Foundation, Inc.,
20 51 Franklin Street - Fifth Floor, Boston, MA 02110-1301, USA. */
21
22 #include "sysdep.h"
23 #include "libiberty.h"
24 #include "bfd.h"
25 #include "bfdlink.h"
26 #include "libbfd.h"
27 #include "elf-bfd.h"
28 #include "elf/spu.h"
29 #include "elf32-spu.h"
30
31 /* We use RELA style relocs. Don't define USE_REL. */
32
33 static bfd_reloc_status_type spu_elf_rel9 (bfd *, arelent *, asymbol *,
34 void *, asection *,
35 bfd *, char **);
36
37 /* Values of type 'enum elf_spu_reloc_type' are used to index this
38 array, so it must be declared in the order of that type. */
39
40 static reloc_howto_type elf_howto_table[] = {
41 HOWTO (R_SPU_NONE, 0, 0, 0, FALSE, 0, complain_overflow_dont,
42 bfd_elf_generic_reloc, "SPU_NONE",
43 FALSE, 0, 0x00000000, FALSE),
44 HOWTO (R_SPU_ADDR10, 4, 2, 10, FALSE, 14, complain_overflow_bitfield,
45 bfd_elf_generic_reloc, "SPU_ADDR10",
46 FALSE, 0, 0x00ffc000, FALSE),
47 HOWTO (R_SPU_ADDR16, 2, 2, 16, FALSE, 7, complain_overflow_bitfield,
48 bfd_elf_generic_reloc, "SPU_ADDR16",
49 FALSE, 0, 0x007fff80, FALSE),
50 HOWTO (R_SPU_ADDR16_HI, 16, 2, 16, FALSE, 7, complain_overflow_bitfield,
51 bfd_elf_generic_reloc, "SPU_ADDR16_HI",
52 FALSE, 0, 0x007fff80, FALSE),
53 HOWTO (R_SPU_ADDR16_LO, 0, 2, 16, FALSE, 7, complain_overflow_dont,
54 bfd_elf_generic_reloc, "SPU_ADDR16_LO",
55 FALSE, 0, 0x007fff80, FALSE),
56 HOWTO (R_SPU_ADDR18, 0, 2, 18, FALSE, 7, complain_overflow_bitfield,
57 bfd_elf_generic_reloc, "SPU_ADDR18",
58 FALSE, 0, 0x01ffff80, FALSE),
59 HOWTO (R_SPU_ADDR32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
60 bfd_elf_generic_reloc, "SPU_ADDR32",
61 FALSE, 0, 0xffffffff, FALSE),
62 HOWTO (R_SPU_REL16, 2, 2, 16, TRUE, 7, complain_overflow_bitfield,
63 bfd_elf_generic_reloc, "SPU_REL16",
64 FALSE, 0, 0x007fff80, TRUE),
65 HOWTO (R_SPU_ADDR7, 0, 2, 7, FALSE, 14, complain_overflow_dont,
66 bfd_elf_generic_reloc, "SPU_ADDR7",
67 FALSE, 0, 0x001fc000, FALSE),
68 HOWTO (R_SPU_REL9, 2, 2, 9, TRUE, 0, complain_overflow_signed,
69 spu_elf_rel9, "SPU_REL9",
70 FALSE, 0, 0x0180007f, TRUE),
71 HOWTO (R_SPU_REL9I, 2, 2, 9, TRUE, 0, complain_overflow_signed,
72 spu_elf_rel9, "SPU_REL9I",
73 FALSE, 0, 0x0000c07f, TRUE),
74 HOWTO (R_SPU_ADDR10I, 0, 2, 10, FALSE, 14, complain_overflow_signed,
75 bfd_elf_generic_reloc, "SPU_ADDR10I",
76 FALSE, 0, 0x00ffc000, FALSE),
77 HOWTO (R_SPU_ADDR16I, 0, 2, 16, FALSE, 7, complain_overflow_signed,
78 bfd_elf_generic_reloc, "SPU_ADDR16I",
79 FALSE, 0, 0x007fff80, FALSE),
80 HOWTO (R_SPU_REL32, 0, 2, 32, TRUE, 0, complain_overflow_dont,
81 bfd_elf_generic_reloc, "SPU_REL32",
82 FALSE, 0, 0xffffffff, TRUE),
83 HOWTO (R_SPU_ADDR16X, 0, 2, 16, FALSE, 7, complain_overflow_bitfield,
84 bfd_elf_generic_reloc, "SPU_ADDR16X",
85 FALSE, 0, 0x007fff80, FALSE),
86 HOWTO (R_SPU_PPU32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
87 bfd_elf_generic_reloc, "SPU_PPU32",
88 FALSE, 0, 0xffffffff, FALSE),
89 HOWTO (R_SPU_PPU64, 0, 4, 64, FALSE, 0, complain_overflow_dont,
90 bfd_elf_generic_reloc, "SPU_PPU64",
91 FALSE, 0, -1, FALSE),
92 HOWTO (R_SPU_ADD_PIC, 0, 0, 0, FALSE, 0, complain_overflow_dont,
93 bfd_elf_generic_reloc, "SPU_ADD_PIC",
94 FALSE, 0, 0x00000000, FALSE),
95 };
96
97 static struct bfd_elf_special_section const spu_elf_special_sections[] = {
98 { "._ea", 4, 0, SHT_PROGBITS, SHF_WRITE },
99 { ".toe", 4, 0, SHT_NOBITS, SHF_ALLOC },
100 { NULL, 0, 0, 0, 0 }
101 };
102
103 static enum elf_spu_reloc_type
104 spu_elf_bfd_to_reloc_type (bfd_reloc_code_real_type code)
105 {
106 switch (code)
107 {
108 default:
109 return R_SPU_NONE;
110 case BFD_RELOC_SPU_IMM10W:
111 return R_SPU_ADDR10;
112 case BFD_RELOC_SPU_IMM16W:
113 return R_SPU_ADDR16;
114 case BFD_RELOC_SPU_LO16:
115 return R_SPU_ADDR16_LO;
116 case BFD_RELOC_SPU_HI16:
117 return R_SPU_ADDR16_HI;
118 case BFD_RELOC_SPU_IMM18:
119 return R_SPU_ADDR18;
120 case BFD_RELOC_SPU_PCREL16:
121 return R_SPU_REL16;
122 case BFD_RELOC_SPU_IMM7:
123 return R_SPU_ADDR7;
124 case BFD_RELOC_SPU_IMM8:
125 return R_SPU_NONE;
126 case BFD_RELOC_SPU_PCREL9a:
127 return R_SPU_REL9;
128 case BFD_RELOC_SPU_PCREL9b:
129 return R_SPU_REL9I;
130 case BFD_RELOC_SPU_IMM10:
131 return R_SPU_ADDR10I;
132 case BFD_RELOC_SPU_IMM16:
133 return R_SPU_ADDR16I;
134 case BFD_RELOC_32:
135 return R_SPU_ADDR32;
136 case BFD_RELOC_32_PCREL:
137 return R_SPU_REL32;
138 case BFD_RELOC_SPU_PPU32:
139 return R_SPU_PPU32;
140 case BFD_RELOC_SPU_PPU64:
141 return R_SPU_PPU64;
142 case BFD_RELOC_SPU_ADD_PIC:
143 return R_SPU_ADD_PIC;
144 }
145 }
146
147 static void
148 spu_elf_info_to_howto (bfd *abfd ATTRIBUTE_UNUSED,
149 arelent *cache_ptr,
150 Elf_Internal_Rela *dst)
151 {
152 enum elf_spu_reloc_type r_type;
153
154 r_type = (enum elf_spu_reloc_type) ELF32_R_TYPE (dst->r_info);
155 BFD_ASSERT (r_type < R_SPU_max);
156 cache_ptr->howto = &elf_howto_table[(int) r_type];
157 }
158
159 static reloc_howto_type *
160 spu_elf_reloc_type_lookup (bfd *abfd ATTRIBUTE_UNUSED,
161 bfd_reloc_code_real_type code)
162 {
163 enum elf_spu_reloc_type r_type = spu_elf_bfd_to_reloc_type (code);
164
165 if (r_type == R_SPU_NONE)
166 return NULL;
167
168 return elf_howto_table + r_type;
169 }
170
171 static reloc_howto_type *
172 spu_elf_reloc_name_lookup (bfd *abfd ATTRIBUTE_UNUSED,
173 const char *r_name)
174 {
175 unsigned int i;
176
177 for (i = 0; i < sizeof (elf_howto_table) / sizeof (elf_howto_table[0]); i++)
178 if (elf_howto_table[i].name != NULL
179 && strcasecmp (elf_howto_table[i].name, r_name) == 0)
180 return &elf_howto_table[i];
181
182 return NULL;
183 }
184
185 /* Apply R_SPU_REL9 and R_SPU_REL9I relocs. */
186
187 static bfd_reloc_status_type
188 spu_elf_rel9 (bfd *abfd, arelent *reloc_entry, asymbol *symbol,
189 void *data, asection *input_section,
190 bfd *output_bfd, char **error_message)
191 {
192 bfd_size_type octets;
193 bfd_vma val;
194 long insn;
195
196 /* If this is a relocatable link (output_bfd test tells us), just
197 call the generic function. Any adjustment will be done at final
198 link time. */
199 if (output_bfd != NULL)
200 return bfd_elf_generic_reloc (abfd, reloc_entry, symbol, data,
201 input_section, output_bfd, error_message);
202
203 if (reloc_entry->address > bfd_get_section_limit (abfd, input_section))
204 return bfd_reloc_outofrange;
205 octets = reloc_entry->address * bfd_octets_per_byte (abfd);
206
207 /* Get symbol value. */
208 val = 0;
209 if (!bfd_is_com_section (symbol->section))
210 val = symbol->value;
211 if (symbol->section->output_section)
212 val += symbol->section->output_section->vma;
213
214 val += reloc_entry->addend;
215
216 /* Make it pc-relative. */
217 val -= input_section->output_section->vma + input_section->output_offset;
218
219 val >>= 2;
220 if (val + 256 >= 512)
221 return bfd_reloc_overflow;
222
223 insn = bfd_get_32 (abfd, (bfd_byte *) data + octets);
224
225 /* Move two high bits of value to REL9I and REL9 position.
226 The mask will take care of selecting the right field. */
227 val = (val & 0x7f) | ((val & 0x180) << 7) | ((val & 0x180) << 16);
228 insn &= ~reloc_entry->howto->dst_mask;
229 insn |= val & reloc_entry->howto->dst_mask;
230 bfd_put_32 (abfd, insn, (bfd_byte *) data + octets);
231 return bfd_reloc_ok;
232 }
233
234 static bfd_boolean
235 spu_elf_new_section_hook (bfd *abfd, asection *sec)
236 {
237 if (!sec->used_by_bfd)
238 {
239 struct _spu_elf_section_data *sdata;
240
241 sdata = bfd_zalloc (abfd, sizeof (*sdata));
242 if (sdata == NULL)
243 return FALSE;
244 sec->used_by_bfd = sdata;
245 }
246
247 return _bfd_elf_new_section_hook (abfd, sec);
248 }
249
250 /* Set up overlay info for executables. */
251
252 static bfd_boolean
253 spu_elf_object_p (bfd *abfd)
254 {
255 if ((abfd->flags & (EXEC_P | DYNAMIC)) != 0)
256 {
257 unsigned int i, num_ovl, num_buf;
258 Elf_Internal_Phdr *phdr = elf_tdata (abfd)->phdr;
259 Elf_Internal_Ehdr *ehdr = elf_elfheader (abfd);
260 Elf_Internal_Phdr *last_phdr = NULL;
261
262 for (num_buf = 0, num_ovl = 0, i = 0; i < ehdr->e_phnum; i++, phdr++)
263 if (phdr->p_type == PT_LOAD && (phdr->p_flags & PF_OVERLAY) != 0)
264 {
265 unsigned int j;
266
267 ++num_ovl;
268 if (last_phdr == NULL
269 || ((last_phdr->p_vaddr ^ phdr->p_vaddr) & 0x3ffff) != 0)
270 ++num_buf;
271 last_phdr = phdr;
272 for (j = 1; j < elf_numsections (abfd); j++)
273 {
274 Elf_Internal_Shdr *shdr = elf_elfsections (abfd)[j];
275
276 if (ELF_SECTION_SIZE (shdr, phdr) != 0
277 && ELF_SECTION_IN_SEGMENT (shdr, phdr))
278 {
279 asection *sec = shdr->bfd_section;
280 spu_elf_section_data (sec)->u.o.ovl_index = num_ovl;
281 spu_elf_section_data (sec)->u.o.ovl_buf = num_buf;
282 }
283 }
284 }
285 }
286 return TRUE;
287 }
288
289 /* Specially mark defined symbols named _EAR_* with BSF_KEEP so that
290 strip --strip-unneeded will not remove them. */
291
292 static void
293 spu_elf_backend_symbol_processing (bfd *abfd ATTRIBUTE_UNUSED, asymbol *sym)
294 {
295 if (sym->name != NULL
296 && sym->section != bfd_abs_section_ptr
297 && strncmp (sym->name, "_EAR_", 5) == 0)
298 sym->flags |= BSF_KEEP;
299 }
300
301 /* SPU ELF linker hash table. */
302
303 struct spu_link_hash_table
304 {
305 struct elf_link_hash_table elf;
306
307 struct spu_elf_params *params;
308
309 /* Shortcuts to overlay sections. */
310 asection *ovtab;
311 asection *init;
312 asection *toe;
313 asection **ovl_sec;
314
315 /* Count of stubs in each overlay section. */
316 unsigned int *stub_count;
317
318 /* The stub section for each overlay section. */
319 asection **stub_sec;
320
321 struct elf_link_hash_entry *ovly_entry[2];
322
323 /* Number of overlay buffers. */
324 unsigned int num_buf;
325
326 /* Total number of overlays. */
327 unsigned int num_overlays;
328
329 /* For soft icache. */
330 unsigned int line_size_log2;
331 unsigned int num_lines_log2;
332 unsigned int fromelem_size_log2;
333
334 /* How much memory we have. */
335 unsigned int local_store;
336
337 /* Count of overlay stubs needed in non-overlay area. */
338 unsigned int non_ovly_stub;
339
340 /* Pointer to the fixup section */
341 asection *sfixup;
342
343 /* Set on error. */
344 unsigned int stub_err : 1;
345 };
346
347 /* Hijack the generic got fields for overlay stub accounting. */
348
349 struct got_entry
350 {
351 struct got_entry *next;
352 unsigned int ovl;
353 union {
354 bfd_vma addend;
355 bfd_vma br_addr;
356 };
357 bfd_vma stub_addr;
358 };
359
360 #define spu_hash_table(p) \
361 (elf_hash_table_id ((struct elf_link_hash_table *) ((p)->hash)) \
362 == SPU_ELF_DATA ? ((struct spu_link_hash_table *) ((p)->hash)) : NULL)
363
364 struct call_info
365 {
366 struct function_info *fun;
367 struct call_info *next;
368 unsigned int count;
369 unsigned int max_depth;
370 unsigned int is_tail : 1;
371 unsigned int is_pasted : 1;
372 unsigned int broken_cycle : 1;
373 unsigned int priority : 13;
374 };
375
376 struct function_info
377 {
378 /* List of functions called. Also branches to hot/cold part of
379 function. */
380 struct call_info *call_list;
381 /* For hot/cold part of function, point to owner. */
382 struct function_info *start;
383 /* Symbol at start of function. */
384 union {
385 Elf_Internal_Sym *sym;
386 struct elf_link_hash_entry *h;
387 } u;
388 /* Function section. */
389 asection *sec;
390 asection *rodata;
391 /* Where last called from, and number of sections called from. */
392 asection *last_caller;
393 unsigned int call_count;
394 /* Address range of (this part of) function. */
395 bfd_vma lo, hi;
396 /* Offset where we found a store of lr, or -1 if none found. */
397 bfd_vma lr_store;
398 /* Offset where we found the stack adjustment insn. */
399 bfd_vma sp_adjust;
400 /* Stack usage. */
401 int stack;
402 /* Distance from root of call tree. Tail and hot/cold branches
403 count as one deeper. We aren't counting stack frames here. */
404 unsigned int depth;
405 /* Set if global symbol. */
406 unsigned int global : 1;
407 /* Set if known to be start of function (as distinct from a hunk
408 in hot/cold section. */
409 unsigned int is_func : 1;
410 /* Set if not a root node. */
411 unsigned int non_root : 1;
412 /* Flags used during call tree traversal. It's cheaper to replicate
413 the visit flags than have one which needs clearing after a traversal. */
414 unsigned int visit1 : 1;
415 unsigned int visit2 : 1;
416 unsigned int marking : 1;
417 unsigned int visit3 : 1;
418 unsigned int visit4 : 1;
419 unsigned int visit5 : 1;
420 unsigned int visit6 : 1;
421 unsigned int visit7 : 1;
422 };
423
424 struct spu_elf_stack_info
425 {
426 int num_fun;
427 int max_fun;
428 /* Variable size array describing functions, one per contiguous
429 address range belonging to a function. */
430 struct function_info fun[1];
431 };
432
433 static struct function_info *find_function (asection *, bfd_vma,
434 struct bfd_link_info *);
435
436 /* Create a spu ELF linker hash table. */
437
438 static struct bfd_link_hash_table *
439 spu_elf_link_hash_table_create (bfd *abfd)
440 {
441 struct spu_link_hash_table *htab;
442
443 htab = bfd_malloc (sizeof (*htab));
444 if (htab == NULL)
445 return NULL;
446
447 if (!_bfd_elf_link_hash_table_init (&htab->elf, abfd,
448 _bfd_elf_link_hash_newfunc,
449 sizeof (struct elf_link_hash_entry),
450 SPU_ELF_DATA))
451 {
452 free (htab);
453 return NULL;
454 }
455
456 memset (&htab->ovtab, 0,
457 sizeof (*htab) - offsetof (struct spu_link_hash_table, ovtab));
458
459 htab->elf.init_got_refcount.refcount = 0;
460 htab->elf.init_got_refcount.glist = NULL;
461 htab->elf.init_got_offset.offset = 0;
462 htab->elf.init_got_offset.glist = NULL;
463 return &htab->elf.root;
464 }
465
466 void
467 spu_elf_setup (struct bfd_link_info *info, struct spu_elf_params *params)
468 {
469 bfd_vma max_branch_log2;
470
471 struct spu_link_hash_table *htab = spu_hash_table (info);
472 htab->params = params;
473 htab->line_size_log2 = bfd_log2 (htab->params->line_size);
474 htab->num_lines_log2 = bfd_log2 (htab->params->num_lines);
475
476 /* For the software i-cache, we provide a "from" list whose size
477 is a power-of-two number of quadwords, big enough to hold one
478 byte per outgoing branch. Compute this number here. */
479 max_branch_log2 = bfd_log2 (htab->params->max_branch);
480 htab->fromelem_size_log2 = max_branch_log2 > 4 ? max_branch_log2 - 4 : 0;
481 }
482
483 /* Find the symbol for the given R_SYMNDX in IBFD and set *HP and *SYMP
484 to (hash, NULL) for global symbols, and (NULL, sym) for locals. Set
485 *SYMSECP to the symbol's section. *LOCSYMSP caches local syms. */
486
487 static bfd_boolean
488 get_sym_h (struct elf_link_hash_entry **hp,
489 Elf_Internal_Sym **symp,
490 asection **symsecp,
491 Elf_Internal_Sym **locsymsp,
492 unsigned long r_symndx,
493 bfd *ibfd)
494 {
495 Elf_Internal_Shdr *symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
496
497 if (r_symndx >= symtab_hdr->sh_info)
498 {
499 struct elf_link_hash_entry **sym_hashes = elf_sym_hashes (ibfd);
500 struct elf_link_hash_entry *h;
501
502 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
503 while (h->root.type == bfd_link_hash_indirect
504 || h->root.type == bfd_link_hash_warning)
505 h = (struct elf_link_hash_entry *) h->root.u.i.link;
506
507 if (hp != NULL)
508 *hp = h;
509
510 if (symp != NULL)
511 *symp = NULL;
512
513 if (symsecp != NULL)
514 {
515 asection *symsec = NULL;
516 if (h->root.type == bfd_link_hash_defined
517 || h->root.type == bfd_link_hash_defweak)
518 symsec = h->root.u.def.section;
519 *symsecp = symsec;
520 }
521 }
522 else
523 {
524 Elf_Internal_Sym *sym;
525 Elf_Internal_Sym *locsyms = *locsymsp;
526
527 if (locsyms == NULL)
528 {
529 locsyms = (Elf_Internal_Sym *) symtab_hdr->contents;
530 if (locsyms == NULL)
531 locsyms = bfd_elf_get_elf_syms (ibfd, symtab_hdr,
532 symtab_hdr->sh_info,
533 0, NULL, NULL, NULL);
534 if (locsyms == NULL)
535 return FALSE;
536 *locsymsp = locsyms;
537 }
538 sym = locsyms + r_symndx;
539
540 if (hp != NULL)
541 *hp = NULL;
542
543 if (symp != NULL)
544 *symp = sym;
545
546 if (symsecp != NULL)
547 *symsecp = bfd_section_from_elf_index (ibfd, sym->st_shndx);
548 }
549
550 return TRUE;
551 }
552
553 /* Create the note section if not already present. This is done early so
554 that the linker maps the sections to the right place in the output. */
555
556 bfd_boolean
557 spu_elf_create_sections (struct bfd_link_info *info)
558 {
559 struct spu_link_hash_table *htab = spu_hash_table (info);
560 bfd *ibfd;
561
562 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
563 if (bfd_get_section_by_name (ibfd, SPU_PTNOTE_SPUNAME) != NULL)
564 break;
565
566 if (ibfd == NULL)
567 {
568 /* Make SPU_PTNOTE_SPUNAME section. */
569 asection *s;
570 size_t name_len;
571 size_t size;
572 bfd_byte *data;
573 flagword flags;
574
575 ibfd = info->input_bfds;
576 flags = SEC_LOAD | SEC_READONLY | SEC_HAS_CONTENTS | SEC_IN_MEMORY;
577 s = bfd_make_section_anyway_with_flags (ibfd, SPU_PTNOTE_SPUNAME, flags);
578 if (s == NULL
579 || !bfd_set_section_alignment (ibfd, s, 4))
580 return FALSE;
581
582 name_len = strlen (bfd_get_filename (info->output_bfd)) + 1;
583 size = 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4);
584 size += (name_len + 3) & -4;
585
586 if (!bfd_set_section_size (ibfd, s, size))
587 return FALSE;
588
589 data = bfd_zalloc (ibfd, size);
590 if (data == NULL)
591 return FALSE;
592
593 bfd_put_32 (ibfd, sizeof (SPU_PLUGIN_NAME), data + 0);
594 bfd_put_32 (ibfd, name_len, data + 4);
595 bfd_put_32 (ibfd, 1, data + 8);
596 memcpy (data + 12, SPU_PLUGIN_NAME, sizeof (SPU_PLUGIN_NAME));
597 memcpy (data + 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4),
598 bfd_get_filename (info->output_bfd), name_len);
599 s->contents = data;
600 }
601
602 if (htab->params->emit_fixups)
603 {
604 asection *s;
605 flagword flags;
606
607 if (htab->elf.dynobj == NULL)
608 htab->elf.dynobj = ibfd;
609 ibfd = htab->elf.dynobj;
610 flags = (SEC_LOAD | SEC_ALLOC | SEC_READONLY | SEC_HAS_CONTENTS
611 | SEC_IN_MEMORY | SEC_LINKER_CREATED);
612 s = bfd_make_section_anyway_with_flags (ibfd, ".fixup", flags);
613 if (s == NULL || !bfd_set_section_alignment (ibfd, s, 2))
614 return FALSE;
615 htab->sfixup = s;
616 }
617
618 return TRUE;
619 }
620
621 /* qsort predicate to sort sections by vma. */
622
623 static int
624 sort_sections (const void *a, const void *b)
625 {
626 const asection *const *s1 = a;
627 const asection *const *s2 = b;
628 bfd_signed_vma delta = (*s1)->vma - (*s2)->vma;
629
630 if (delta != 0)
631 return delta < 0 ? -1 : 1;
632
633 return (*s1)->index - (*s2)->index;
634 }
635
636 /* Identify overlays in the output bfd, and number them.
637 Returns 0 on error, 1 if no overlays, 2 if overlays. */
638
639 int
640 spu_elf_find_overlays (struct bfd_link_info *info)
641 {
642 struct spu_link_hash_table *htab = spu_hash_table (info);
643 asection **alloc_sec;
644 unsigned int i, n, ovl_index, num_buf;
645 asection *s;
646 bfd_vma ovl_end;
647 static const char *const entry_names[2][2] = {
648 { "__ovly_load", "__icache_br_handler" },
649 { "__ovly_return", "__icache_call_handler" }
650 };
651
652 if (info->output_bfd->section_count < 2)
653 return 1;
654
655 alloc_sec
656 = bfd_malloc (info->output_bfd->section_count * sizeof (*alloc_sec));
657 if (alloc_sec == NULL)
658 return 0;
659
660 /* Pick out all the alloced sections. */
661 for (n = 0, s = info->output_bfd->sections; s != NULL; s = s->next)
662 if ((s->flags & SEC_ALLOC) != 0
663 && (s->flags & (SEC_LOAD | SEC_THREAD_LOCAL)) != SEC_THREAD_LOCAL
664 && s->size != 0)
665 alloc_sec[n++] = s;
666
667 if (n == 0)
668 {
669 free (alloc_sec);
670 return 1;
671 }
672
673 /* Sort them by vma. */
674 qsort (alloc_sec, n, sizeof (*alloc_sec), sort_sections);
675
676 ovl_end = alloc_sec[0]->vma + alloc_sec[0]->size;
677 if (htab->params->ovly_flavour == ovly_soft_icache)
678 {
679 unsigned int prev_buf = 0, set_id = 0;
680
681 /* Look for an overlapping vma to find the first overlay section. */
682 bfd_vma vma_start = 0;
683
684 for (i = 1; i < n; i++)
685 {
686 s = alloc_sec[i];
687 if (s->vma < ovl_end)
688 {
689 asection *s0 = alloc_sec[i - 1];
690 vma_start = s0->vma;
691 ovl_end = (s0->vma
692 + ((bfd_vma) 1
693 << (htab->num_lines_log2 + htab->line_size_log2)));
694 --i;
695 break;
696 }
697 else
698 ovl_end = s->vma + s->size;
699 }
700
701 /* Now find any sections within the cache area. */
702 for (ovl_index = 0, num_buf = 0; i < n; i++)
703 {
704 s = alloc_sec[i];
705 if (s->vma >= ovl_end)
706 break;
707
708 /* A section in an overlay area called .ovl.init is not
709 an overlay, in the sense that it might be loaded in
710 by the overlay manager, but rather the initial
711 section contents for the overlay buffer. */
712 if (strncmp (s->name, ".ovl.init", 9) != 0)
713 {
714 num_buf = ((s->vma - vma_start) >> htab->line_size_log2) + 1;
715 set_id = (num_buf == prev_buf)? set_id + 1 : 0;
716 prev_buf = num_buf;
717
718 if ((s->vma - vma_start) & (htab->params->line_size - 1))
719 {
720 info->callbacks->einfo (_("%X%P: overlay section %A "
721 "does not start on a cache line.\n"),
722 s);
723 bfd_set_error (bfd_error_bad_value);
724 return 0;
725 }
726 else if (s->size > htab->params->line_size)
727 {
728 info->callbacks->einfo (_("%X%P: overlay section %A "
729 "is larger than a cache line.\n"),
730 s);
731 bfd_set_error (bfd_error_bad_value);
732 return 0;
733 }
734
735 alloc_sec[ovl_index++] = s;
736 spu_elf_section_data (s)->u.o.ovl_index
737 = (set_id << htab->num_lines_log2) + num_buf;
738 spu_elf_section_data (s)->u.o.ovl_buf = num_buf;
739 }
740 }
741
742 /* Ensure there are no more overlay sections. */
743 for ( ; i < n; i++)
744 {
745 s = alloc_sec[i];
746 if (s->vma < ovl_end)
747 {
748 info->callbacks->einfo (_("%X%P: overlay section %A "
749 "is not in cache area.\n"),
750 alloc_sec[i-1]);
751 bfd_set_error (bfd_error_bad_value);
752 return 0;
753 }
754 else
755 ovl_end = s->vma + s->size;
756 }
757 }
758 else
759 {
760 /* Look for overlapping vmas. Any with overlap must be overlays.
761 Count them. Also count the number of overlay regions. */
762 for (ovl_index = 0, num_buf = 0, i = 1; i < n; i++)
763 {
764 s = alloc_sec[i];
765 if (s->vma < ovl_end)
766 {
767 asection *s0 = alloc_sec[i - 1];
768
769 if (spu_elf_section_data (s0)->u.o.ovl_index == 0)
770 {
771 ++num_buf;
772 if (strncmp (s0->name, ".ovl.init", 9) != 0)
773 {
774 alloc_sec[ovl_index] = s0;
775 spu_elf_section_data (s0)->u.o.ovl_index = ++ovl_index;
776 spu_elf_section_data (s0)->u.o.ovl_buf = num_buf;
777 }
778 else
779 ovl_end = s->vma + s->size;
780 }
781 if (strncmp (s->name, ".ovl.init", 9) != 0)
782 {
783 alloc_sec[ovl_index] = s;
784 spu_elf_section_data (s)->u.o.ovl_index = ++ovl_index;
785 spu_elf_section_data (s)->u.o.ovl_buf = num_buf;
786 if (s0->vma != s->vma)
787 {
788 info->callbacks->einfo (_("%X%P: overlay sections %A "
789 "and %A do not start at the "
790 "same address.\n"),
791 s0, s);
792 bfd_set_error (bfd_error_bad_value);
793 return 0;
794 }
795 if (ovl_end < s->vma + s->size)
796 ovl_end = s->vma + s->size;
797 }
798 }
799 else
800 ovl_end = s->vma + s->size;
801 }
802 }
803
804 htab->num_overlays = ovl_index;
805 htab->num_buf = num_buf;
806 htab->ovl_sec = alloc_sec;
807
808 if (ovl_index == 0)
809 return 1;
810
811 for (i = 0; i < 2; i++)
812 {
813 const char *name;
814 struct elf_link_hash_entry *h;
815
816 name = entry_names[i][htab->params->ovly_flavour];
817 h = elf_link_hash_lookup (&htab->elf, name, TRUE, FALSE, FALSE);
818 if (h == NULL)
819 return 0;
820
821 if (h->root.type == bfd_link_hash_new)
822 {
823 h->root.type = bfd_link_hash_undefined;
824 h->ref_regular = 1;
825 h->ref_regular_nonweak = 1;
826 h->non_elf = 0;
827 }
828 htab->ovly_entry[i] = h;
829 }
830
831 return 2;
832 }
833
834 /* Non-zero to use bra in overlay stubs rather than br. */
835 #define BRA_STUBS 0
836
837 #define BRA 0x30000000
838 #define BRASL 0x31000000
839 #define BR 0x32000000
840 #define BRSL 0x33000000
841 #define NOP 0x40200000
842 #define LNOP 0x00200000
843 #define ILA 0x42000000
844
845 /* Return true for all relative and absolute branch instructions.
846 bra 00110000 0..
847 brasl 00110001 0..
848 br 00110010 0..
849 brsl 00110011 0..
850 brz 00100000 0..
851 brnz 00100001 0..
852 brhz 00100010 0..
853 brhnz 00100011 0.. */
854
855 static bfd_boolean
856 is_branch (const unsigned char *insn)
857 {
858 return (insn[0] & 0xec) == 0x20 && (insn[1] & 0x80) == 0;
859 }
860
861 /* Return true for all indirect branch instructions.
862 bi 00110101 000
863 bisl 00110101 001
864 iret 00110101 010
865 bisled 00110101 011
866 biz 00100101 000
867 binz 00100101 001
868 bihz 00100101 010
869 bihnz 00100101 011 */
870
871 static bfd_boolean
872 is_indirect_branch (const unsigned char *insn)
873 {
874 return (insn[0] & 0xef) == 0x25 && (insn[1] & 0x80) == 0;
875 }
876
877 /* Return true for branch hint instructions.
878 hbra 0001000..
879 hbrr 0001001.. */
880
881 static bfd_boolean
882 is_hint (const unsigned char *insn)
883 {
884 return (insn[0] & 0xfc) == 0x10;
885 }
886
887 /* True if INPUT_SECTION might need overlay stubs. */
888
889 static bfd_boolean
890 maybe_needs_stubs (asection *input_section)
891 {
892 /* No stubs for debug sections and suchlike. */
893 if ((input_section->flags & SEC_ALLOC) == 0)
894 return FALSE;
895
896 /* No stubs for link-once sections that will be discarded. */
897 if (input_section->output_section == bfd_abs_section_ptr)
898 return FALSE;
899
900 /* Don't create stubs for .eh_frame references. */
901 if (strcmp (input_section->name, ".eh_frame") == 0)
902 return FALSE;
903
904 return TRUE;
905 }
906
907 enum _stub_type
908 {
909 no_stub,
910 call_ovl_stub,
911 br000_ovl_stub,
912 br001_ovl_stub,
913 br010_ovl_stub,
914 br011_ovl_stub,
915 br100_ovl_stub,
916 br101_ovl_stub,
917 br110_ovl_stub,
918 br111_ovl_stub,
919 nonovl_stub,
920 stub_error
921 };
922
923 /* Return non-zero if this reloc symbol should go via an overlay stub.
924 Return 2 if the stub must be in non-overlay area. */
925
926 static enum _stub_type
927 needs_ovl_stub (struct elf_link_hash_entry *h,
928 Elf_Internal_Sym *sym,
929 asection *sym_sec,
930 asection *input_section,
931 Elf_Internal_Rela *irela,
932 bfd_byte *contents,
933 struct bfd_link_info *info)
934 {
935 struct spu_link_hash_table *htab = spu_hash_table (info);
936 enum elf_spu_reloc_type r_type;
937 unsigned int sym_type;
938 bfd_boolean branch, hint, call;
939 enum _stub_type ret = no_stub;
940 bfd_byte insn[4];
941
942 if (sym_sec == NULL
943 || sym_sec->output_section == bfd_abs_section_ptr
944 || spu_elf_section_data (sym_sec->output_section) == NULL)
945 return ret;
946
947 if (h != NULL)
948 {
949 /* Ensure no stubs for user supplied overlay manager syms. */
950 if (h == htab->ovly_entry[0] || h == htab->ovly_entry[1])
951 return ret;
952
953 /* setjmp always goes via an overlay stub, because then the return
954 and hence the longjmp goes via __ovly_return. That magically
955 makes setjmp/longjmp between overlays work. */
956 if (strncmp (h->root.root.string, "setjmp", 6) == 0
957 && (h->root.root.string[6] == '\0' || h->root.root.string[6] == '@'))
958 ret = call_ovl_stub;
959 }
960
961 if (h != NULL)
962 sym_type = h->type;
963 else
964 sym_type = ELF_ST_TYPE (sym->st_info);
965
966 r_type = ELF32_R_TYPE (irela->r_info);
967 branch = FALSE;
968 hint = FALSE;
969 call = FALSE;
970 if (r_type == R_SPU_REL16 || r_type == R_SPU_ADDR16)
971 {
972 if (contents == NULL)
973 {
974 contents = insn;
975 if (!bfd_get_section_contents (input_section->owner,
976 input_section,
977 contents,
978 irela->r_offset, 4))
979 return stub_error;
980 }
981 else
982 contents += irela->r_offset;
983
984 branch = is_branch (contents);
985 hint = is_hint (contents);
986 if (branch || hint)
987 {
988 call = (contents[0] & 0xfd) == 0x31;
989 if (call
990 && sym_type != STT_FUNC
991 && contents != insn)
992 {
993 /* It's common for people to write assembly and forget
994 to give function symbols the right type. Handle
995 calls to such symbols, but warn so that (hopefully)
996 people will fix their code. We need the symbol
997 type to be correct to distinguish function pointer
998 initialisation from other pointer initialisations. */
999 const char *sym_name;
1000
1001 if (h != NULL)
1002 sym_name = h->root.root.string;
1003 else
1004 {
1005 Elf_Internal_Shdr *symtab_hdr;
1006 symtab_hdr = &elf_tdata (input_section->owner)->symtab_hdr;
1007 sym_name = bfd_elf_sym_name (input_section->owner,
1008 symtab_hdr,
1009 sym,
1010 sym_sec);
1011 }
1012 (*_bfd_error_handler) (_("warning: call to non-function"
1013 " symbol %s defined in %B"),
1014 sym_sec->owner, sym_name);
1015
1016 }
1017 }
1018 }
1019
1020 if ((!branch && htab->params->ovly_flavour == ovly_soft_icache)
1021 || (sym_type != STT_FUNC
1022 && !(branch || hint)
1023 && (sym_sec->flags & SEC_CODE) == 0))
1024 return no_stub;
1025
1026 /* Usually, symbols in non-overlay sections don't need stubs. */
1027 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index == 0
1028 && !htab->params->non_overlay_stubs)
1029 return ret;
1030
1031 /* A reference from some other section to a symbol in an overlay
1032 section needs a stub. */
1033 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index
1034 != spu_elf_section_data (input_section->output_section)->u.o.ovl_index)
1035 {
1036 unsigned int lrlive = 0;
1037 if (branch)
1038 lrlive = (contents[1] & 0x70) >> 4;
1039
1040 if (!lrlive && (call || sym_type == STT_FUNC))
1041 ret = call_ovl_stub;
1042 else
1043 ret = br000_ovl_stub + lrlive;
1044 }
1045
1046 /* If this insn isn't a branch then we are possibly taking the
1047 address of a function and passing it out somehow. Soft-icache code
1048 always generates inline code to do indirect branches. */
1049 if (!(branch || hint)
1050 && sym_type == STT_FUNC
1051 && htab->params->ovly_flavour != ovly_soft_icache)
1052 ret = nonovl_stub;
1053
1054 return ret;
1055 }
1056
1057 static bfd_boolean
1058 count_stub (struct spu_link_hash_table *htab,
1059 bfd *ibfd,
1060 asection *isec,
1061 enum _stub_type stub_type,
1062 struct elf_link_hash_entry *h,
1063 const Elf_Internal_Rela *irela)
1064 {
1065 unsigned int ovl = 0;
1066 struct got_entry *g, **head;
1067 bfd_vma addend;
1068
1069 /* If this instruction is a branch or call, we need a stub
1070 for it. One stub per function per overlay.
1071 If it isn't a branch, then we are taking the address of
1072 this function so need a stub in the non-overlay area
1073 for it. One stub per function. */
1074 if (stub_type != nonovl_stub)
1075 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
1076
1077 if (h != NULL)
1078 head = &h->got.glist;
1079 else
1080 {
1081 if (elf_local_got_ents (ibfd) == NULL)
1082 {
1083 bfd_size_type amt = (elf_tdata (ibfd)->symtab_hdr.sh_info
1084 * sizeof (*elf_local_got_ents (ibfd)));
1085 elf_local_got_ents (ibfd) = bfd_zmalloc (amt);
1086 if (elf_local_got_ents (ibfd) == NULL)
1087 return FALSE;
1088 }
1089 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
1090 }
1091
1092 if (htab->params->ovly_flavour == ovly_soft_icache)
1093 {
1094 htab->stub_count[ovl] += 1;
1095 return TRUE;
1096 }
1097
1098 addend = 0;
1099 if (irela != NULL)
1100 addend = irela->r_addend;
1101
1102 if (ovl == 0)
1103 {
1104 struct got_entry *gnext;
1105
1106 for (g = *head; g != NULL; g = g->next)
1107 if (g->addend == addend && g->ovl == 0)
1108 break;
1109
1110 if (g == NULL)
1111 {
1112 /* Need a new non-overlay area stub. Zap other stubs. */
1113 for (g = *head; g != NULL; g = gnext)
1114 {
1115 gnext = g->next;
1116 if (g->addend == addend)
1117 {
1118 htab->stub_count[g->ovl] -= 1;
1119 free (g);
1120 }
1121 }
1122 }
1123 }
1124 else
1125 {
1126 for (g = *head; g != NULL; g = g->next)
1127 if (g->addend == addend && (g->ovl == ovl || g->ovl == 0))
1128 break;
1129 }
1130
1131 if (g == NULL)
1132 {
1133 g = bfd_malloc (sizeof *g);
1134 if (g == NULL)
1135 return FALSE;
1136 g->ovl = ovl;
1137 g->addend = addend;
1138 g->stub_addr = (bfd_vma) -1;
1139 g->next = *head;
1140 *head = g;
1141
1142 htab->stub_count[ovl] += 1;
1143 }
1144
1145 return TRUE;
1146 }
1147
1148 /* Support two sizes of overlay stubs, a slower more compact stub of two
1149 intructions, and a faster stub of four instructions.
1150 Soft-icache stubs are four or eight words. */
1151
1152 static unsigned int
1153 ovl_stub_size (struct spu_elf_params *params)
1154 {
1155 return 16 << params->ovly_flavour >> params->compact_stub;
1156 }
1157
1158 static unsigned int
1159 ovl_stub_size_log2 (struct spu_elf_params *params)
1160 {
1161 return 4 + params->ovly_flavour - params->compact_stub;
1162 }
1163
1164 /* Two instruction overlay stubs look like:
1165
1166 brsl $75,__ovly_load
1167 .word target_ovl_and_address
1168
1169 ovl_and_address is a word with the overlay number in the top 14 bits
1170 and local store address in the bottom 18 bits.
1171
1172 Four instruction overlay stubs look like:
1173
1174 ila $78,ovl_number
1175 lnop
1176 ila $79,target_address
1177 br __ovly_load
1178
1179 Software icache stubs are:
1180
1181 .word target_index
1182 .word target_ia;
1183 .word lrlive_branchlocalstoreaddr;
1184 brasl $75,__icache_br_handler
1185 .quad xor_pattern
1186 */
1187
1188 static bfd_boolean
1189 build_stub (struct bfd_link_info *info,
1190 bfd *ibfd,
1191 asection *isec,
1192 enum _stub_type stub_type,
1193 struct elf_link_hash_entry *h,
1194 const Elf_Internal_Rela *irela,
1195 bfd_vma dest,
1196 asection *dest_sec)
1197 {
1198 struct spu_link_hash_table *htab = spu_hash_table (info);
1199 unsigned int ovl, dest_ovl, set_id;
1200 struct got_entry *g, **head;
1201 asection *sec;
1202 bfd_vma addend, from, to, br_dest, patt;
1203 unsigned int lrlive;
1204
1205 ovl = 0;
1206 if (stub_type != nonovl_stub)
1207 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
1208
1209 if (h != NULL)
1210 head = &h->got.glist;
1211 else
1212 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
1213
1214 addend = 0;
1215 if (irela != NULL)
1216 addend = irela->r_addend;
1217
1218 if (htab->params->ovly_flavour == ovly_soft_icache)
1219 {
1220 g = bfd_malloc (sizeof *g);
1221 if (g == NULL)
1222 return FALSE;
1223 g->ovl = ovl;
1224 g->br_addr = 0;
1225 if (irela != NULL)
1226 g->br_addr = (irela->r_offset
1227 + isec->output_offset
1228 + isec->output_section->vma);
1229 g->next = *head;
1230 *head = g;
1231 }
1232 else
1233 {
1234 for (g = *head; g != NULL; g = g->next)
1235 if (g->addend == addend && (g->ovl == ovl || g->ovl == 0))
1236 break;
1237 if (g == NULL)
1238 abort ();
1239
1240 if (g->ovl == 0 && ovl != 0)
1241 return TRUE;
1242
1243 if (g->stub_addr != (bfd_vma) -1)
1244 return TRUE;
1245 }
1246
1247 sec = htab->stub_sec[ovl];
1248 dest += dest_sec->output_offset + dest_sec->output_section->vma;
1249 from = sec->size + sec->output_offset + sec->output_section->vma;
1250 g->stub_addr = from;
1251 to = (htab->ovly_entry[0]->root.u.def.value
1252 + htab->ovly_entry[0]->root.u.def.section->output_offset
1253 + htab->ovly_entry[0]->root.u.def.section->output_section->vma);
1254
1255 if (((dest | to | from) & 3) != 0)
1256 {
1257 htab->stub_err = 1;
1258 return FALSE;
1259 }
1260 dest_ovl = spu_elf_section_data (dest_sec->output_section)->u.o.ovl_index;
1261
1262 if (htab->params->ovly_flavour == ovly_normal
1263 && !htab->params->compact_stub)
1264 {
1265 bfd_put_32 (sec->owner, ILA + ((dest_ovl << 7) & 0x01ffff80) + 78,
1266 sec->contents + sec->size);
1267 bfd_put_32 (sec->owner, LNOP,
1268 sec->contents + sec->size + 4);
1269 bfd_put_32 (sec->owner, ILA + ((dest << 7) & 0x01ffff80) + 79,
1270 sec->contents + sec->size + 8);
1271 if (!BRA_STUBS)
1272 bfd_put_32 (sec->owner, BR + (((to - (from + 12)) << 5) & 0x007fff80),
1273 sec->contents + sec->size + 12);
1274 else
1275 bfd_put_32 (sec->owner, BRA + ((to << 5) & 0x007fff80),
1276 sec->contents + sec->size + 12);
1277 }
1278 else if (htab->params->ovly_flavour == ovly_normal
1279 && htab->params->compact_stub)
1280 {
1281 if (!BRA_STUBS)
1282 bfd_put_32 (sec->owner, BRSL + (((to - from) << 5) & 0x007fff80) + 75,
1283 sec->contents + sec->size);
1284 else
1285 bfd_put_32 (sec->owner, BRASL + ((to << 5) & 0x007fff80) + 75,
1286 sec->contents + sec->size);
1287 bfd_put_32 (sec->owner, (dest & 0x3ffff) | (dest_ovl << 18),
1288 sec->contents + sec->size + 4);
1289 }
1290 else if (htab->params->ovly_flavour == ovly_soft_icache
1291 && htab->params->compact_stub)
1292 {
1293 lrlive = 0;
1294 if (stub_type == nonovl_stub)
1295 ;
1296 else if (stub_type == call_ovl_stub)
1297 /* A brsl makes lr live and *(*sp+16) is live.
1298 Tail calls have the same liveness. */
1299 lrlive = 5;
1300 else if (!htab->params->lrlive_analysis)
1301 /* Assume stack frame and lr save. */
1302 lrlive = 1;
1303 else if (irela != NULL)
1304 {
1305 /* Analyse branch instructions. */
1306 struct function_info *caller;
1307 bfd_vma off;
1308
1309 caller = find_function (isec, irela->r_offset, info);
1310 if (caller->start == NULL)
1311 off = irela->r_offset;
1312 else
1313 {
1314 struct function_info *found = NULL;
1315
1316 /* Find the earliest piece of this function that
1317 has frame adjusting instructions. We might
1318 see dynamic frame adjustment (eg. for alloca)
1319 in some later piece, but functions using
1320 alloca always set up a frame earlier. Frame
1321 setup instructions are always in one piece. */
1322 if (caller->lr_store != (bfd_vma) -1
1323 || caller->sp_adjust != (bfd_vma) -1)
1324 found = caller;
1325 while (caller->start != NULL)
1326 {
1327 caller = caller->start;
1328 if (caller->lr_store != (bfd_vma) -1
1329 || caller->sp_adjust != (bfd_vma) -1)
1330 found = caller;
1331 }
1332 if (found != NULL)
1333 caller = found;
1334 off = (bfd_vma) -1;
1335 }
1336
1337 if (off > caller->sp_adjust)
1338 {
1339 if (off > caller->lr_store)
1340 /* Only *(*sp+16) is live. */
1341 lrlive = 1;
1342 else
1343 /* If no lr save, then we must be in a
1344 leaf function with a frame.
1345 lr is still live. */
1346 lrlive = 4;
1347 }
1348 else if (off > caller->lr_store)
1349 {
1350 /* Between lr save and stack adjust. */
1351 lrlive = 3;
1352 /* This should never happen since prologues won't
1353 be split here. */
1354 BFD_ASSERT (0);
1355 }
1356 else
1357 /* On entry to function. */
1358 lrlive = 5;
1359
1360 if (stub_type != br000_ovl_stub
1361 && lrlive != stub_type - br000_ovl_stub)
1362 info->callbacks->einfo (_("%A:0x%v lrlive .brinfo (%u) differs "
1363 "from analysis (%u)\n"),
1364 isec, irela->r_offset, lrlive,
1365 stub_type - br000_ovl_stub);
1366 }
1367
1368 /* If given lrlive info via .brinfo, use it. */
1369 if (stub_type > br000_ovl_stub)
1370 lrlive = stub_type - br000_ovl_stub;
1371
1372 if (ovl == 0)
1373 to = (htab->ovly_entry[1]->root.u.def.value
1374 + htab->ovly_entry[1]->root.u.def.section->output_offset
1375 + htab->ovly_entry[1]->root.u.def.section->output_section->vma);
1376
1377 /* The branch that uses this stub goes to stub_addr + 4. We'll
1378 set up an xor pattern that can be used by the icache manager
1379 to modify this branch to go directly to its destination. */
1380 g->stub_addr += 4;
1381 br_dest = g->stub_addr;
1382 if (irela == NULL)
1383 {
1384 /* Except in the case of _SPUEAR_ stubs, the branch in
1385 question is the one in the stub itself. */
1386 BFD_ASSERT (stub_type == nonovl_stub);
1387 g->br_addr = g->stub_addr;
1388 br_dest = to;
1389 }
1390
1391 set_id = ((dest_ovl - 1) >> htab->num_lines_log2) + 1;
1392 bfd_put_32 (sec->owner, (set_id << 18) | (dest & 0x3ffff),
1393 sec->contents + sec->size);
1394 bfd_put_32 (sec->owner, BRASL + ((to << 5) & 0x007fff80) + 75,
1395 sec->contents + sec->size + 4);
1396 bfd_put_32 (sec->owner, (lrlive << 29) | (g->br_addr & 0x3ffff),
1397 sec->contents + sec->size + 8);
1398 patt = dest ^ br_dest;
1399 if (irela != NULL && ELF32_R_TYPE (irela->r_info) == R_SPU_REL16)
1400 patt = (dest - g->br_addr) ^ (br_dest - g->br_addr);
1401 bfd_put_32 (sec->owner, (patt << 5) & 0x007fff80,
1402 sec->contents + sec->size + 12);
1403
1404 if (ovl == 0)
1405 /* Extra space for linked list entries. */
1406 sec->size += 16;
1407 }
1408 else
1409 abort ();
1410
1411 sec->size += ovl_stub_size (htab->params);
1412
1413 if (htab->params->emit_stub_syms)
1414 {
1415 size_t len;
1416 char *name;
1417 int add;
1418
1419 len = 8 + sizeof (".ovl_call.") - 1;
1420 if (h != NULL)
1421 len += strlen (h->root.root.string);
1422 else
1423 len += 8 + 1 + 8;
1424 add = 0;
1425 if (irela != NULL)
1426 add = (int) irela->r_addend & 0xffffffff;
1427 if (add != 0)
1428 len += 1 + 8;
1429 name = bfd_malloc (len + 1);
1430 if (name == NULL)
1431 return FALSE;
1432
1433 sprintf (name, "%08x.ovl_call.", g->ovl);
1434 if (h != NULL)
1435 strcpy (name + 8 + sizeof (".ovl_call.") - 1, h->root.root.string);
1436 else
1437 sprintf (name + 8 + sizeof (".ovl_call.") - 1, "%x:%x",
1438 dest_sec->id & 0xffffffff,
1439 (int) ELF32_R_SYM (irela->r_info) & 0xffffffff);
1440 if (add != 0)
1441 sprintf (name + len - 9, "+%x", add);
1442
1443 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
1444 free (name);
1445 if (h == NULL)
1446 return FALSE;
1447 if (h->root.type == bfd_link_hash_new)
1448 {
1449 h->root.type = bfd_link_hash_defined;
1450 h->root.u.def.section = sec;
1451 h->size = ovl_stub_size (htab->params);
1452 h->root.u.def.value = sec->size - h->size;
1453 h->type = STT_FUNC;
1454 h->ref_regular = 1;
1455 h->def_regular = 1;
1456 h->ref_regular_nonweak = 1;
1457 h->forced_local = 1;
1458 h->non_elf = 0;
1459 }
1460 }
1461
1462 return TRUE;
1463 }
1464
1465 /* Called via elf_link_hash_traverse to allocate stubs for any _SPUEAR_
1466 symbols. */
1467
1468 static bfd_boolean
1469 allocate_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
1470 {
1471 /* Symbols starting with _SPUEAR_ need a stub because they may be
1472 invoked by the PPU. */
1473 struct bfd_link_info *info = inf;
1474 struct spu_link_hash_table *htab = spu_hash_table (info);
1475 asection *sym_sec;
1476
1477 if ((h->root.type == bfd_link_hash_defined
1478 || h->root.type == bfd_link_hash_defweak)
1479 && h->def_regular
1480 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0
1481 && (sym_sec = h->root.u.def.section) != NULL
1482 && sym_sec->output_section != bfd_abs_section_ptr
1483 && spu_elf_section_data (sym_sec->output_section) != NULL
1484 && (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index != 0
1485 || htab->params->non_overlay_stubs))
1486 {
1487 return count_stub (htab, NULL, NULL, nonovl_stub, h, NULL);
1488 }
1489
1490 return TRUE;
1491 }
1492
1493 static bfd_boolean
1494 build_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
1495 {
1496 /* Symbols starting with _SPUEAR_ need a stub because they may be
1497 invoked by the PPU. */
1498 struct bfd_link_info *info = inf;
1499 struct spu_link_hash_table *htab = spu_hash_table (info);
1500 asection *sym_sec;
1501
1502 if ((h->root.type == bfd_link_hash_defined
1503 || h->root.type == bfd_link_hash_defweak)
1504 && h->def_regular
1505 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0
1506 && (sym_sec = h->root.u.def.section) != NULL
1507 && sym_sec->output_section != bfd_abs_section_ptr
1508 && spu_elf_section_data (sym_sec->output_section) != NULL
1509 && (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index != 0
1510 || htab->params->non_overlay_stubs))
1511 {
1512 return build_stub (info, NULL, NULL, nonovl_stub, h, NULL,
1513 h->root.u.def.value, sym_sec);
1514 }
1515
1516 return TRUE;
1517 }
1518
1519 /* Size or build stubs. */
1520
1521 static bfd_boolean
1522 process_stubs (struct bfd_link_info *info, bfd_boolean build)
1523 {
1524 struct spu_link_hash_table *htab = spu_hash_table (info);
1525 bfd *ibfd;
1526
1527 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
1528 {
1529 extern const bfd_target bfd_elf32_spu_vec;
1530 Elf_Internal_Shdr *symtab_hdr;
1531 asection *isec;
1532 Elf_Internal_Sym *local_syms = NULL;
1533
1534 if (ibfd->xvec != &bfd_elf32_spu_vec)
1535 continue;
1536
1537 /* We'll need the symbol table in a second. */
1538 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
1539 if (symtab_hdr->sh_info == 0)
1540 continue;
1541
1542 /* Walk over each section attached to the input bfd. */
1543 for (isec = ibfd->sections; isec != NULL; isec = isec->next)
1544 {
1545 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
1546
1547 /* If there aren't any relocs, then there's nothing more to do. */
1548 if ((isec->flags & SEC_RELOC) == 0
1549 || isec->reloc_count == 0)
1550 continue;
1551
1552 if (!maybe_needs_stubs (isec))
1553 continue;
1554
1555 /* Get the relocs. */
1556 internal_relocs = _bfd_elf_link_read_relocs (ibfd, isec, NULL, NULL,
1557 info->keep_memory);
1558 if (internal_relocs == NULL)
1559 goto error_ret_free_local;
1560
1561 /* Now examine each relocation. */
1562 irela = internal_relocs;
1563 irelaend = irela + isec->reloc_count;
1564 for (; irela < irelaend; irela++)
1565 {
1566 enum elf_spu_reloc_type r_type;
1567 unsigned int r_indx;
1568 asection *sym_sec;
1569 Elf_Internal_Sym *sym;
1570 struct elf_link_hash_entry *h;
1571 enum _stub_type stub_type;
1572
1573 r_type = ELF32_R_TYPE (irela->r_info);
1574 r_indx = ELF32_R_SYM (irela->r_info);
1575
1576 if (r_type >= R_SPU_max)
1577 {
1578 bfd_set_error (bfd_error_bad_value);
1579 error_ret_free_internal:
1580 if (elf_section_data (isec)->relocs != internal_relocs)
1581 free (internal_relocs);
1582 error_ret_free_local:
1583 if (local_syms != NULL
1584 && (symtab_hdr->contents
1585 != (unsigned char *) local_syms))
1586 free (local_syms);
1587 return FALSE;
1588 }
1589
1590 /* Determine the reloc target section. */
1591 if (!get_sym_h (&h, &sym, &sym_sec, &local_syms, r_indx, ibfd))
1592 goto error_ret_free_internal;
1593
1594 stub_type = needs_ovl_stub (h, sym, sym_sec, isec, irela,
1595 NULL, info);
1596 if (stub_type == no_stub)
1597 continue;
1598 else if (stub_type == stub_error)
1599 goto error_ret_free_internal;
1600
1601 if (htab->stub_count == NULL)
1602 {
1603 bfd_size_type amt;
1604 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_count);
1605 htab->stub_count = bfd_zmalloc (amt);
1606 if (htab->stub_count == NULL)
1607 goto error_ret_free_internal;
1608 }
1609
1610 if (!build)
1611 {
1612 if (!count_stub (htab, ibfd, isec, stub_type, h, irela))
1613 goto error_ret_free_internal;
1614 }
1615 else
1616 {
1617 bfd_vma dest;
1618
1619 if (h != NULL)
1620 dest = h->root.u.def.value;
1621 else
1622 dest = sym->st_value;
1623 dest += irela->r_addend;
1624 if (!build_stub (info, ibfd, isec, stub_type, h, irela,
1625 dest, sym_sec))
1626 goto error_ret_free_internal;
1627 }
1628 }
1629
1630 /* We're done with the internal relocs, free them. */
1631 if (elf_section_data (isec)->relocs != internal_relocs)
1632 free (internal_relocs);
1633 }
1634
1635 if (local_syms != NULL
1636 && symtab_hdr->contents != (unsigned char *) local_syms)
1637 {
1638 if (!info->keep_memory)
1639 free (local_syms);
1640 else
1641 symtab_hdr->contents = (unsigned char *) local_syms;
1642 }
1643 }
1644
1645 return TRUE;
1646 }
1647
1648 /* Allocate space for overlay call and return stubs.
1649 Return 0 on error, 1 if no overlays, 2 otherwise. */
1650
1651 int
1652 spu_elf_size_stubs (struct bfd_link_info *info)
1653 {
1654 struct spu_link_hash_table *htab;
1655 bfd *ibfd;
1656 bfd_size_type amt;
1657 flagword flags;
1658 unsigned int i;
1659 asection *stub;
1660
1661 if (!process_stubs (info, FALSE))
1662 return 0;
1663
1664 htab = spu_hash_table (info);
1665 elf_link_hash_traverse (&htab->elf, allocate_spuear_stubs, info);
1666 if (htab->stub_err)
1667 return 0;
1668
1669 ibfd = info->input_bfds;
1670 if (htab->stub_count != NULL)
1671 {
1672 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_sec);
1673 htab->stub_sec = bfd_zmalloc (amt);
1674 if (htab->stub_sec == NULL)
1675 return 0;
1676
1677 flags = (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_READONLY
1678 | SEC_HAS_CONTENTS | SEC_IN_MEMORY);
1679 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1680 htab->stub_sec[0] = stub;
1681 if (stub == NULL
1682 || !bfd_set_section_alignment (ibfd, stub,
1683 ovl_stub_size_log2 (htab->params)))
1684 return 0;
1685 stub->size = htab->stub_count[0] * ovl_stub_size (htab->params);
1686 if (htab->params->ovly_flavour == ovly_soft_icache)
1687 /* Extra space for linked list entries. */
1688 stub->size += htab->stub_count[0] * 16;
1689
1690 for (i = 0; i < htab->num_overlays; ++i)
1691 {
1692 asection *osec = htab->ovl_sec[i];
1693 unsigned int ovl = spu_elf_section_data (osec)->u.o.ovl_index;
1694 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1695 htab->stub_sec[ovl] = stub;
1696 if (stub == NULL
1697 || !bfd_set_section_alignment (ibfd, stub,
1698 ovl_stub_size_log2 (htab->params)))
1699 return 0;
1700 stub->size = htab->stub_count[ovl] * ovl_stub_size (htab->params);
1701 }
1702 }
1703
1704 if (htab->params->ovly_flavour == ovly_soft_icache)
1705 {
1706 /* Space for icache manager tables.
1707 a) Tag array, one quadword per cache line.
1708 b) Rewrite "to" list, one quadword per cache line.
1709 c) Rewrite "from" list, one byte per outgoing branch (rounded up to
1710 a power-of-two number of full quadwords) per cache line. */
1711
1712 flags = SEC_ALLOC;
1713 htab->ovtab = bfd_make_section_anyway_with_flags (ibfd, ".ovtab", flags);
1714 if (htab->ovtab == NULL
1715 || !bfd_set_section_alignment (ibfd, htab->ovtab, 4))
1716 return 0;
1717
1718 htab->ovtab->size = (16 + 16 + (16 << htab->fromelem_size_log2))
1719 << htab->num_lines_log2;
1720
1721 flags = SEC_ALLOC | SEC_LOAD | SEC_HAS_CONTENTS | SEC_IN_MEMORY;
1722 htab->init = bfd_make_section_anyway_with_flags (ibfd, ".ovini", flags);
1723 if (htab->init == NULL
1724 || !bfd_set_section_alignment (ibfd, htab->init, 4))
1725 return 0;
1726
1727 htab->init->size = 16;
1728 }
1729 else if (htab->stub_count == NULL)
1730 return 1;
1731 else
1732 {
1733 /* htab->ovtab consists of two arrays.
1734 . struct {
1735 . u32 vma;
1736 . u32 size;
1737 . u32 file_off;
1738 . u32 buf;
1739 . } _ovly_table[];
1740 .
1741 . struct {
1742 . u32 mapped;
1743 . } _ovly_buf_table[];
1744 . */
1745
1746 flags = SEC_ALLOC | SEC_LOAD | SEC_HAS_CONTENTS | SEC_IN_MEMORY;
1747 htab->ovtab = bfd_make_section_anyway_with_flags (ibfd, ".ovtab", flags);
1748 if (htab->ovtab == NULL
1749 || !bfd_set_section_alignment (ibfd, htab->ovtab, 4))
1750 return 0;
1751
1752 htab->ovtab->size = htab->num_overlays * 16 + 16 + htab->num_buf * 4;
1753 }
1754
1755 htab->toe = bfd_make_section_anyway_with_flags (ibfd, ".toe", SEC_ALLOC);
1756 if (htab->toe == NULL
1757 || !bfd_set_section_alignment (ibfd, htab->toe, 4))
1758 return 0;
1759 htab->toe->size = 16;
1760
1761 return 2;
1762 }
1763
1764 /* Called from ld to place overlay manager data sections. This is done
1765 after the overlay manager itself is loaded, mainly so that the
1766 linker's htab->init section is placed after any other .ovl.init
1767 sections. */
1768
1769 void
1770 spu_elf_place_overlay_data (struct bfd_link_info *info)
1771 {
1772 struct spu_link_hash_table *htab = spu_hash_table (info);
1773 unsigned int i;
1774
1775 if (htab->stub_sec != NULL)
1776 {
1777 (*htab->params->place_spu_section) (htab->stub_sec[0], NULL, ".text");
1778
1779 for (i = 0; i < htab->num_overlays; ++i)
1780 {
1781 asection *osec = htab->ovl_sec[i];
1782 unsigned int ovl = spu_elf_section_data (osec)->u.o.ovl_index;
1783 (*htab->params->place_spu_section) (htab->stub_sec[ovl], osec, NULL);
1784 }
1785 }
1786
1787 if (htab->params->ovly_flavour == ovly_soft_icache)
1788 (*htab->params->place_spu_section) (htab->init, NULL, ".ovl.init");
1789
1790 if (htab->ovtab != NULL)
1791 {
1792 const char *ovout = ".data";
1793 if (htab->params->ovly_flavour == ovly_soft_icache)
1794 ovout = ".bss";
1795 (*htab->params->place_spu_section) (htab->ovtab, NULL, ovout);
1796 }
1797
1798 if (htab->toe != NULL)
1799 (*htab->params->place_spu_section) (htab->toe, NULL, ".toe");
1800 }
1801
1802 /* Functions to handle embedded spu_ovl.o object. */
1803
1804 static void *
1805 ovl_mgr_open (struct bfd *nbfd ATTRIBUTE_UNUSED, void *stream)
1806 {
1807 return stream;
1808 }
1809
1810 static file_ptr
1811 ovl_mgr_pread (struct bfd *abfd ATTRIBUTE_UNUSED,
1812 void *stream,
1813 void *buf,
1814 file_ptr nbytes,
1815 file_ptr offset)
1816 {
1817 struct _ovl_stream *os;
1818 size_t count;
1819 size_t max;
1820
1821 os = (struct _ovl_stream *) stream;
1822 max = (const char *) os->end - (const char *) os->start;
1823
1824 if ((ufile_ptr) offset >= max)
1825 return 0;
1826
1827 count = nbytes;
1828 if (count > max - offset)
1829 count = max - offset;
1830
1831 memcpy (buf, (const char *) os->start + offset, count);
1832 return count;
1833 }
1834
1835 bfd_boolean
1836 spu_elf_open_builtin_lib (bfd **ovl_bfd, const struct _ovl_stream *stream)
1837 {
1838 *ovl_bfd = bfd_openr_iovec ("builtin ovl_mgr",
1839 "elf32-spu",
1840 ovl_mgr_open,
1841 (void *) stream,
1842 ovl_mgr_pread,
1843 NULL,
1844 NULL);
1845 return *ovl_bfd != NULL;
1846 }
1847
1848 static unsigned int
1849 overlay_index (asection *sec)
1850 {
1851 if (sec == NULL
1852 || sec->output_section == bfd_abs_section_ptr)
1853 return 0;
1854 return spu_elf_section_data (sec->output_section)->u.o.ovl_index;
1855 }
1856
1857 /* Define an STT_OBJECT symbol. */
1858
1859 static struct elf_link_hash_entry *
1860 define_ovtab_symbol (struct spu_link_hash_table *htab, const char *name)
1861 {
1862 struct elf_link_hash_entry *h;
1863
1864 h = elf_link_hash_lookup (&htab->elf, name, TRUE, FALSE, FALSE);
1865 if (h == NULL)
1866 return NULL;
1867
1868 if (h->root.type != bfd_link_hash_defined
1869 || !h->def_regular)
1870 {
1871 h->root.type = bfd_link_hash_defined;
1872 h->root.u.def.section = htab->ovtab;
1873 h->type = STT_OBJECT;
1874 h->ref_regular = 1;
1875 h->def_regular = 1;
1876 h->ref_regular_nonweak = 1;
1877 h->non_elf = 0;
1878 }
1879 else if (h->root.u.def.section->owner != NULL)
1880 {
1881 (*_bfd_error_handler) (_("%B is not allowed to define %s"),
1882 h->root.u.def.section->owner,
1883 h->root.root.string);
1884 bfd_set_error (bfd_error_bad_value);
1885 return NULL;
1886 }
1887 else
1888 {
1889 (*_bfd_error_handler) (_("you are not allowed to define %s in a script"),
1890 h->root.root.string);
1891 bfd_set_error (bfd_error_bad_value);
1892 return NULL;
1893 }
1894
1895 return h;
1896 }
1897
1898 /* Fill in all stubs and the overlay tables. */
1899
1900 static bfd_boolean
1901 spu_elf_build_stubs (struct bfd_link_info *info)
1902 {
1903 struct spu_link_hash_table *htab = spu_hash_table (info);
1904 struct elf_link_hash_entry *h;
1905 bfd_byte *p;
1906 asection *s;
1907 bfd *obfd;
1908 unsigned int i;
1909
1910 if (htab->num_overlays != 0)
1911 {
1912 for (i = 0; i < 2; i++)
1913 {
1914 h = htab->ovly_entry[i];
1915 if (h != NULL
1916 && (h->root.type == bfd_link_hash_defined
1917 || h->root.type == bfd_link_hash_defweak)
1918 && h->def_regular)
1919 {
1920 s = h->root.u.def.section->output_section;
1921 if (spu_elf_section_data (s)->u.o.ovl_index)
1922 {
1923 (*_bfd_error_handler) (_("%s in overlay section"),
1924 h->root.root.string);
1925 bfd_set_error (bfd_error_bad_value);
1926 return FALSE;
1927 }
1928 }
1929 }
1930 }
1931
1932 if (htab->stub_sec != NULL)
1933 {
1934 for (i = 0; i <= htab->num_overlays; i++)
1935 if (htab->stub_sec[i]->size != 0)
1936 {
1937 htab->stub_sec[i]->contents = bfd_zalloc (htab->stub_sec[i]->owner,
1938 htab->stub_sec[i]->size);
1939 if (htab->stub_sec[i]->contents == NULL)
1940 return FALSE;
1941 htab->stub_sec[i]->rawsize = htab->stub_sec[i]->size;
1942 htab->stub_sec[i]->size = 0;
1943 }
1944
1945 /* Fill in all the stubs. */
1946 process_stubs (info, TRUE);
1947 if (!htab->stub_err)
1948 elf_link_hash_traverse (&htab->elf, build_spuear_stubs, info);
1949
1950 if (htab->stub_err)
1951 {
1952 (*_bfd_error_handler) (_("overlay stub relocation overflow"));
1953 bfd_set_error (bfd_error_bad_value);
1954 return FALSE;
1955 }
1956
1957 for (i = 0; i <= htab->num_overlays; i++)
1958 {
1959 if (htab->stub_sec[i]->size != htab->stub_sec[i]->rawsize)
1960 {
1961 (*_bfd_error_handler) (_("stubs don't match calculated size"));
1962 bfd_set_error (bfd_error_bad_value);
1963 return FALSE;
1964 }
1965 htab->stub_sec[i]->rawsize = 0;
1966 }
1967 }
1968
1969 if (htab->ovtab == NULL || htab->ovtab->size == 0)
1970 return TRUE;
1971
1972 htab->ovtab->contents = bfd_zalloc (htab->ovtab->owner, htab->ovtab->size);
1973 if (htab->ovtab->contents == NULL)
1974 return FALSE;
1975
1976 p = htab->ovtab->contents;
1977 if (htab->params->ovly_flavour == ovly_soft_icache)
1978 {
1979 bfd_vma off;
1980
1981 h = define_ovtab_symbol (htab, "__icache_tag_array");
1982 if (h == NULL)
1983 return FALSE;
1984 h->root.u.def.value = 0;
1985 h->size = 16 << htab->num_lines_log2;
1986 off = h->size;
1987
1988 h = define_ovtab_symbol (htab, "__icache_tag_array_size");
1989 if (h == NULL)
1990 return FALSE;
1991 h->root.u.def.value = 16 << htab->num_lines_log2;
1992 h->root.u.def.section = bfd_abs_section_ptr;
1993
1994 h = define_ovtab_symbol (htab, "__icache_rewrite_to");
1995 if (h == NULL)
1996 return FALSE;
1997 h->root.u.def.value = off;
1998 h->size = 16 << htab->num_lines_log2;
1999 off += h->size;
2000
2001 h = define_ovtab_symbol (htab, "__icache_rewrite_to_size");
2002 if (h == NULL)
2003 return FALSE;
2004 h->root.u.def.value = 16 << htab->num_lines_log2;
2005 h->root.u.def.section = bfd_abs_section_ptr;
2006
2007 h = define_ovtab_symbol (htab, "__icache_rewrite_from");
2008 if (h == NULL)
2009 return FALSE;
2010 h->root.u.def.value = off;
2011 h->size = 16 << (htab->fromelem_size_log2 + htab->num_lines_log2);
2012 off += h->size;
2013
2014 h = define_ovtab_symbol (htab, "__icache_rewrite_from_size");
2015 if (h == NULL)
2016 return FALSE;
2017 h->root.u.def.value = 16 << (htab->fromelem_size_log2
2018 + htab->num_lines_log2);
2019 h->root.u.def.section = bfd_abs_section_ptr;
2020
2021 h = define_ovtab_symbol (htab, "__icache_log2_fromelemsize");
2022 if (h == NULL)
2023 return FALSE;
2024 h->root.u.def.value = htab->fromelem_size_log2;
2025 h->root.u.def.section = bfd_abs_section_ptr;
2026
2027 h = define_ovtab_symbol (htab, "__icache_base");
2028 if (h == NULL)
2029 return FALSE;
2030 h->root.u.def.value = htab->ovl_sec[0]->vma;
2031 h->root.u.def.section = bfd_abs_section_ptr;
2032 h->size = htab->num_buf << htab->line_size_log2;
2033
2034 h = define_ovtab_symbol (htab, "__icache_linesize");
2035 if (h == NULL)
2036 return FALSE;
2037 h->root.u.def.value = 1 << htab->line_size_log2;
2038 h->root.u.def.section = bfd_abs_section_ptr;
2039
2040 h = define_ovtab_symbol (htab, "__icache_log2_linesize");
2041 if (h == NULL)
2042 return FALSE;
2043 h->root.u.def.value = htab->line_size_log2;
2044 h->root.u.def.section = bfd_abs_section_ptr;
2045
2046 h = define_ovtab_symbol (htab, "__icache_neg_log2_linesize");
2047 if (h == NULL)
2048 return FALSE;
2049 h->root.u.def.value = -htab->line_size_log2;
2050 h->root.u.def.section = bfd_abs_section_ptr;
2051
2052 h = define_ovtab_symbol (htab, "__icache_cachesize");
2053 if (h == NULL)
2054 return FALSE;
2055 h->root.u.def.value = 1 << (htab->num_lines_log2 + htab->line_size_log2);
2056 h->root.u.def.section = bfd_abs_section_ptr;
2057
2058 h = define_ovtab_symbol (htab, "__icache_log2_cachesize");
2059 if (h == NULL)
2060 return FALSE;
2061 h->root.u.def.value = htab->num_lines_log2 + htab->line_size_log2;
2062 h->root.u.def.section = bfd_abs_section_ptr;
2063
2064 h = define_ovtab_symbol (htab, "__icache_neg_log2_cachesize");
2065 if (h == NULL)
2066 return FALSE;
2067 h->root.u.def.value = -(htab->num_lines_log2 + htab->line_size_log2);
2068 h->root.u.def.section = bfd_abs_section_ptr;
2069
2070 if (htab->init != NULL && htab->init->size != 0)
2071 {
2072 htab->init->contents = bfd_zalloc (htab->init->owner,
2073 htab->init->size);
2074 if (htab->init->contents == NULL)
2075 return FALSE;
2076
2077 h = define_ovtab_symbol (htab, "__icache_fileoff");
2078 if (h == NULL)
2079 return FALSE;
2080 h->root.u.def.value = 0;
2081 h->root.u.def.section = htab->init;
2082 h->size = 8;
2083 }
2084 }
2085 else
2086 {
2087 /* Write out _ovly_table. */
2088 /* set low bit of .size to mark non-overlay area as present. */
2089 p[7] = 1;
2090 obfd = htab->ovtab->output_section->owner;
2091 for (s = obfd->sections; s != NULL; s = s->next)
2092 {
2093 unsigned int ovl_index = spu_elf_section_data (s)->u.o.ovl_index;
2094
2095 if (ovl_index != 0)
2096 {
2097 unsigned long off = ovl_index * 16;
2098 unsigned int ovl_buf = spu_elf_section_data (s)->u.o.ovl_buf;
2099
2100 bfd_put_32 (htab->ovtab->owner, s->vma, p + off);
2101 bfd_put_32 (htab->ovtab->owner, (s->size + 15) & -16,
2102 p + off + 4);
2103 /* file_off written later in spu_elf_modify_program_headers. */
2104 bfd_put_32 (htab->ovtab->owner, ovl_buf, p + off + 12);
2105 }
2106 }
2107
2108 h = define_ovtab_symbol (htab, "_ovly_table");
2109 if (h == NULL)
2110 return FALSE;
2111 h->root.u.def.value = 16;
2112 h->size = htab->num_overlays * 16;
2113
2114 h = define_ovtab_symbol (htab, "_ovly_table_end");
2115 if (h == NULL)
2116 return FALSE;
2117 h->root.u.def.value = htab->num_overlays * 16 + 16;
2118 h->size = 0;
2119
2120 h = define_ovtab_symbol (htab, "_ovly_buf_table");
2121 if (h == NULL)
2122 return FALSE;
2123 h->root.u.def.value = htab->num_overlays * 16 + 16;
2124 h->size = htab->num_buf * 4;
2125
2126 h = define_ovtab_symbol (htab, "_ovly_buf_table_end");
2127 if (h == NULL)
2128 return FALSE;
2129 h->root.u.def.value = htab->num_overlays * 16 + 16 + htab->num_buf * 4;
2130 h->size = 0;
2131 }
2132
2133 h = define_ovtab_symbol (htab, "_EAR_");
2134 if (h == NULL)
2135 return FALSE;
2136 h->root.u.def.section = htab->toe;
2137 h->root.u.def.value = 0;
2138 h->size = 16;
2139
2140 return TRUE;
2141 }
2142
2143 /* Check that all loadable section VMAs lie in the range
2144 LO .. HI inclusive, and stash some parameters for --auto-overlay. */
2145
2146 asection *
2147 spu_elf_check_vma (struct bfd_link_info *info)
2148 {
2149 struct elf_segment_map *m;
2150 unsigned int i;
2151 struct spu_link_hash_table *htab = spu_hash_table (info);
2152 bfd *abfd = info->output_bfd;
2153 bfd_vma hi = htab->params->local_store_hi;
2154 bfd_vma lo = htab->params->local_store_lo;
2155
2156 htab->local_store = hi + 1 - lo;
2157
2158 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
2159 if (m->p_type == PT_LOAD)
2160 for (i = 0; i < m->count; i++)
2161 if (m->sections[i]->size != 0
2162 && (m->sections[i]->vma < lo
2163 || m->sections[i]->vma > hi
2164 || m->sections[i]->vma + m->sections[i]->size - 1 > hi))
2165 return m->sections[i];
2166
2167 return NULL;
2168 }
2169
2170 /* OFFSET in SEC (presumably) is the beginning of a function prologue.
2171 Search for stack adjusting insns, and return the sp delta.
2172 If a store of lr is found save the instruction offset to *LR_STORE.
2173 If a stack adjusting instruction is found, save that offset to
2174 *SP_ADJUST. */
2175
2176 static int
2177 find_function_stack_adjust (asection *sec,
2178 bfd_vma offset,
2179 bfd_vma *lr_store,
2180 bfd_vma *sp_adjust)
2181 {
2182 int reg[128];
2183
2184 memset (reg, 0, sizeof (reg));
2185 for ( ; offset + 4 <= sec->size; offset += 4)
2186 {
2187 unsigned char buf[4];
2188 int rt, ra;
2189 int imm;
2190
2191 /* Assume no relocs on stack adjusing insns. */
2192 if (!bfd_get_section_contents (sec->owner, sec, buf, offset, 4))
2193 break;
2194
2195 rt = buf[3] & 0x7f;
2196 ra = ((buf[2] & 0x3f) << 1) | (buf[3] >> 7);
2197
2198 if (buf[0] == 0x24 /* stqd */)
2199 {
2200 if (rt == 0 /* lr */ && ra == 1 /* sp */)
2201 *lr_store = offset;
2202 continue;
2203 }
2204
2205 /* Partly decoded immediate field. */
2206 imm = (buf[1] << 9) | (buf[2] << 1) | (buf[3] >> 7);
2207
2208 if (buf[0] == 0x1c /* ai */)
2209 {
2210 imm >>= 7;
2211 imm = (imm ^ 0x200) - 0x200;
2212 reg[rt] = reg[ra] + imm;
2213
2214 if (rt == 1 /* sp */)
2215 {
2216 if (reg[rt] > 0)
2217 break;
2218 *sp_adjust = offset;
2219 return reg[rt];
2220 }
2221 }
2222 else if (buf[0] == 0x18 && (buf[1] & 0xe0) == 0 /* a */)
2223 {
2224 int rb = ((buf[1] & 0x1f) << 2) | ((buf[2] & 0xc0) >> 6);
2225
2226 reg[rt] = reg[ra] + reg[rb];
2227 if (rt == 1)
2228 {
2229 if (reg[rt] > 0)
2230 break;
2231 *sp_adjust = offset;
2232 return reg[rt];
2233 }
2234 }
2235 else if (buf[0] == 0x08 && (buf[1] & 0xe0) == 0 /* sf */)
2236 {
2237 int rb = ((buf[1] & 0x1f) << 2) | ((buf[2] & 0xc0) >> 6);
2238
2239 reg[rt] = reg[rb] - reg[ra];
2240 if (rt == 1)
2241 {
2242 if (reg[rt] > 0)
2243 break;
2244 *sp_adjust = offset;
2245 return reg[rt];
2246 }
2247 }
2248 else if ((buf[0] & 0xfc) == 0x40 /* il, ilh, ilhu, ila */)
2249 {
2250 if (buf[0] >= 0x42 /* ila */)
2251 imm |= (buf[0] & 1) << 17;
2252 else
2253 {
2254 imm &= 0xffff;
2255
2256 if (buf[0] == 0x40 /* il */)
2257 {
2258 if ((buf[1] & 0x80) == 0)
2259 continue;
2260 imm = (imm ^ 0x8000) - 0x8000;
2261 }
2262 else if ((buf[1] & 0x80) == 0 /* ilhu */)
2263 imm <<= 16;
2264 }
2265 reg[rt] = imm;
2266 continue;
2267 }
2268 else if (buf[0] == 0x60 && (buf[1] & 0x80) != 0 /* iohl */)
2269 {
2270 reg[rt] |= imm & 0xffff;
2271 continue;
2272 }
2273 else if (buf[0] == 0x04 /* ori */)
2274 {
2275 imm >>= 7;
2276 imm = (imm ^ 0x200) - 0x200;
2277 reg[rt] = reg[ra] | imm;
2278 continue;
2279 }
2280 else if (buf[0] == 0x32 && (buf[1] & 0x80) != 0 /* fsmbi */)
2281 {
2282 reg[rt] = ( ((imm & 0x8000) ? 0xff000000 : 0)
2283 | ((imm & 0x4000) ? 0x00ff0000 : 0)
2284 | ((imm & 0x2000) ? 0x0000ff00 : 0)
2285 | ((imm & 0x1000) ? 0x000000ff : 0));
2286 continue;
2287 }
2288 else if (buf[0] == 0x16 /* andbi */)
2289 {
2290 imm >>= 7;
2291 imm &= 0xff;
2292 imm |= imm << 8;
2293 imm |= imm << 16;
2294 reg[rt] = reg[ra] & imm;
2295 continue;
2296 }
2297 else if (buf[0] == 0x33 && imm == 1 /* brsl .+4 */)
2298 {
2299 /* Used in pic reg load. Say rt is trashed. Won't be used
2300 in stack adjust, but we need to continue past this branch. */
2301 reg[rt] = 0;
2302 continue;
2303 }
2304 else if (is_branch (buf) || is_indirect_branch (buf))
2305 /* If we hit a branch then we must be out of the prologue. */
2306 break;
2307 }
2308
2309 return 0;
2310 }
2311
2312 /* qsort predicate to sort symbols by section and value. */
2313
2314 static Elf_Internal_Sym *sort_syms_syms;
2315 static asection **sort_syms_psecs;
2316
2317 static int
2318 sort_syms (const void *a, const void *b)
2319 {
2320 Elf_Internal_Sym *const *s1 = a;
2321 Elf_Internal_Sym *const *s2 = b;
2322 asection *sec1,*sec2;
2323 bfd_signed_vma delta;
2324
2325 sec1 = sort_syms_psecs[*s1 - sort_syms_syms];
2326 sec2 = sort_syms_psecs[*s2 - sort_syms_syms];
2327
2328 if (sec1 != sec2)
2329 return sec1->index - sec2->index;
2330
2331 delta = (*s1)->st_value - (*s2)->st_value;
2332 if (delta != 0)
2333 return delta < 0 ? -1 : 1;
2334
2335 delta = (*s2)->st_size - (*s1)->st_size;
2336 if (delta != 0)
2337 return delta < 0 ? -1 : 1;
2338
2339 return *s1 < *s2 ? -1 : 1;
2340 }
2341
2342 /* Allocate a struct spu_elf_stack_info with MAX_FUN struct function_info
2343 entries for section SEC. */
2344
2345 static struct spu_elf_stack_info *
2346 alloc_stack_info (asection *sec, int max_fun)
2347 {
2348 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2349 bfd_size_type amt;
2350
2351 amt = sizeof (struct spu_elf_stack_info);
2352 amt += (max_fun - 1) * sizeof (struct function_info);
2353 sec_data->u.i.stack_info = bfd_zmalloc (amt);
2354 if (sec_data->u.i.stack_info != NULL)
2355 sec_data->u.i.stack_info->max_fun = max_fun;
2356 return sec_data->u.i.stack_info;
2357 }
2358
2359 /* Add a new struct function_info describing a (part of a) function
2360 starting at SYM_H. Keep the array sorted by address. */
2361
2362 static struct function_info *
2363 maybe_insert_function (asection *sec,
2364 void *sym_h,
2365 bfd_boolean global,
2366 bfd_boolean is_func)
2367 {
2368 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2369 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2370 int i;
2371 bfd_vma off, size;
2372
2373 if (sinfo == NULL)
2374 {
2375 sinfo = alloc_stack_info (sec, 20);
2376 if (sinfo == NULL)
2377 return NULL;
2378 }
2379
2380 if (!global)
2381 {
2382 Elf_Internal_Sym *sym = sym_h;
2383 off = sym->st_value;
2384 size = sym->st_size;
2385 }
2386 else
2387 {
2388 struct elf_link_hash_entry *h = sym_h;
2389 off = h->root.u.def.value;
2390 size = h->size;
2391 }
2392
2393 for (i = sinfo->num_fun; --i >= 0; )
2394 if (sinfo->fun[i].lo <= off)
2395 break;
2396
2397 if (i >= 0)
2398 {
2399 /* Don't add another entry for an alias, but do update some
2400 info. */
2401 if (sinfo->fun[i].lo == off)
2402 {
2403 /* Prefer globals over local syms. */
2404 if (global && !sinfo->fun[i].global)
2405 {
2406 sinfo->fun[i].global = TRUE;
2407 sinfo->fun[i].u.h = sym_h;
2408 }
2409 if (is_func)
2410 sinfo->fun[i].is_func = TRUE;
2411 return &sinfo->fun[i];
2412 }
2413 /* Ignore a zero-size symbol inside an existing function. */
2414 else if (sinfo->fun[i].hi > off && size == 0)
2415 return &sinfo->fun[i];
2416 }
2417
2418 if (sinfo->num_fun >= sinfo->max_fun)
2419 {
2420 bfd_size_type amt = sizeof (struct spu_elf_stack_info);
2421 bfd_size_type old = amt;
2422
2423 old += (sinfo->max_fun - 1) * sizeof (struct function_info);
2424 sinfo->max_fun += 20 + (sinfo->max_fun >> 1);
2425 amt += (sinfo->max_fun - 1) * sizeof (struct function_info);
2426 sinfo = bfd_realloc (sinfo, amt);
2427 if (sinfo == NULL)
2428 return NULL;
2429 memset ((char *) sinfo + old, 0, amt - old);
2430 sec_data->u.i.stack_info = sinfo;
2431 }
2432
2433 if (++i < sinfo->num_fun)
2434 memmove (&sinfo->fun[i + 1], &sinfo->fun[i],
2435 (sinfo->num_fun - i) * sizeof (sinfo->fun[i]));
2436 sinfo->fun[i].is_func = is_func;
2437 sinfo->fun[i].global = global;
2438 sinfo->fun[i].sec = sec;
2439 if (global)
2440 sinfo->fun[i].u.h = sym_h;
2441 else
2442 sinfo->fun[i].u.sym = sym_h;
2443 sinfo->fun[i].lo = off;
2444 sinfo->fun[i].hi = off + size;
2445 sinfo->fun[i].lr_store = -1;
2446 sinfo->fun[i].sp_adjust = -1;
2447 sinfo->fun[i].stack = -find_function_stack_adjust (sec, off,
2448 &sinfo->fun[i].lr_store,
2449 &sinfo->fun[i].sp_adjust);
2450 sinfo->num_fun += 1;
2451 return &sinfo->fun[i];
2452 }
2453
2454 /* Return the name of FUN. */
2455
2456 static const char *
2457 func_name (struct function_info *fun)
2458 {
2459 asection *sec;
2460 bfd *ibfd;
2461 Elf_Internal_Shdr *symtab_hdr;
2462
2463 while (fun->start != NULL)
2464 fun = fun->start;
2465
2466 if (fun->global)
2467 return fun->u.h->root.root.string;
2468
2469 sec = fun->sec;
2470 if (fun->u.sym->st_name == 0)
2471 {
2472 size_t len = strlen (sec->name);
2473 char *name = bfd_malloc (len + 10);
2474 if (name == NULL)
2475 return "(null)";
2476 sprintf (name, "%s+%lx", sec->name,
2477 (unsigned long) fun->u.sym->st_value & 0xffffffff);
2478 return name;
2479 }
2480 ibfd = sec->owner;
2481 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2482 return bfd_elf_sym_name (ibfd, symtab_hdr, fun->u.sym, sec);
2483 }
2484
2485 /* Read the instruction at OFF in SEC. Return true iff the instruction
2486 is a nop, lnop, or stop 0 (all zero insn). */
2487
2488 static bfd_boolean
2489 is_nop (asection *sec, bfd_vma off)
2490 {
2491 unsigned char insn[4];
2492
2493 if (off + 4 > sec->size
2494 || !bfd_get_section_contents (sec->owner, sec, insn, off, 4))
2495 return FALSE;
2496 if ((insn[0] & 0xbf) == 0 && (insn[1] & 0xe0) == 0x20)
2497 return TRUE;
2498 if (insn[0] == 0 && insn[1] == 0 && insn[2] == 0 && insn[3] == 0)
2499 return TRUE;
2500 return FALSE;
2501 }
2502
2503 /* Extend the range of FUN to cover nop padding up to LIMIT.
2504 Return TRUE iff some instruction other than a NOP was found. */
2505
2506 static bfd_boolean
2507 insns_at_end (struct function_info *fun, bfd_vma limit)
2508 {
2509 bfd_vma off = (fun->hi + 3) & -4;
2510
2511 while (off < limit && is_nop (fun->sec, off))
2512 off += 4;
2513 if (off < limit)
2514 {
2515 fun->hi = off;
2516 return TRUE;
2517 }
2518 fun->hi = limit;
2519 return FALSE;
2520 }
2521
2522 /* Check and fix overlapping function ranges. Return TRUE iff there
2523 are gaps in the current info we have about functions in SEC. */
2524
2525 static bfd_boolean
2526 check_function_ranges (asection *sec, struct bfd_link_info *info)
2527 {
2528 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2529 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2530 int i;
2531 bfd_boolean gaps = FALSE;
2532
2533 if (sinfo == NULL)
2534 return FALSE;
2535
2536 for (i = 1; i < sinfo->num_fun; i++)
2537 if (sinfo->fun[i - 1].hi > sinfo->fun[i].lo)
2538 {
2539 /* Fix overlapping symbols. */
2540 const char *f1 = func_name (&sinfo->fun[i - 1]);
2541 const char *f2 = func_name (&sinfo->fun[i]);
2542
2543 info->callbacks->einfo (_("warning: %s overlaps %s\n"), f1, f2);
2544 sinfo->fun[i - 1].hi = sinfo->fun[i].lo;
2545 }
2546 else if (insns_at_end (&sinfo->fun[i - 1], sinfo->fun[i].lo))
2547 gaps = TRUE;
2548
2549 if (sinfo->num_fun == 0)
2550 gaps = TRUE;
2551 else
2552 {
2553 if (sinfo->fun[0].lo != 0)
2554 gaps = TRUE;
2555 if (sinfo->fun[sinfo->num_fun - 1].hi > sec->size)
2556 {
2557 const char *f1 = func_name (&sinfo->fun[sinfo->num_fun - 1]);
2558
2559 info->callbacks->einfo (_("warning: %s exceeds section size\n"), f1);
2560 sinfo->fun[sinfo->num_fun - 1].hi = sec->size;
2561 }
2562 else if (insns_at_end (&sinfo->fun[sinfo->num_fun - 1], sec->size))
2563 gaps = TRUE;
2564 }
2565 return gaps;
2566 }
2567
2568 /* Search current function info for a function that contains address
2569 OFFSET in section SEC. */
2570
2571 static struct function_info *
2572 find_function (asection *sec, bfd_vma offset, struct bfd_link_info *info)
2573 {
2574 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2575 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2576 int lo, hi, mid;
2577
2578 lo = 0;
2579 hi = sinfo->num_fun;
2580 while (lo < hi)
2581 {
2582 mid = (lo + hi) / 2;
2583 if (offset < sinfo->fun[mid].lo)
2584 hi = mid;
2585 else if (offset >= sinfo->fun[mid].hi)
2586 lo = mid + 1;
2587 else
2588 return &sinfo->fun[mid];
2589 }
2590 info->callbacks->einfo (_("%A:0x%v not found in function table\n"),
2591 sec, offset);
2592 bfd_set_error (bfd_error_bad_value);
2593 return NULL;
2594 }
2595
2596 /* Add CALLEE to CALLER call list if not already present. Return TRUE
2597 if CALLEE was new. If this function return FALSE, CALLEE should
2598 be freed. */
2599
2600 static bfd_boolean
2601 insert_callee (struct function_info *caller, struct call_info *callee)
2602 {
2603 struct call_info **pp, *p;
2604
2605 for (pp = &caller->call_list; (p = *pp) != NULL; pp = &p->next)
2606 if (p->fun == callee->fun)
2607 {
2608 /* Tail calls use less stack than normal calls. Retain entry
2609 for normal call over one for tail call. */
2610 p->is_tail &= callee->is_tail;
2611 if (!p->is_tail)
2612 {
2613 p->fun->start = NULL;
2614 p->fun->is_func = TRUE;
2615 }
2616 p->count += callee->count;
2617 /* Reorder list so most recent call is first. */
2618 *pp = p->next;
2619 p->next = caller->call_list;
2620 caller->call_list = p;
2621 return FALSE;
2622 }
2623 callee->next = caller->call_list;
2624 caller->call_list = callee;
2625 return TRUE;
2626 }
2627
2628 /* Copy CALL and insert the copy into CALLER. */
2629
2630 static bfd_boolean
2631 copy_callee (struct function_info *caller, const struct call_info *call)
2632 {
2633 struct call_info *callee;
2634 callee = bfd_malloc (sizeof (*callee));
2635 if (callee == NULL)
2636 return FALSE;
2637 *callee = *call;
2638 if (!insert_callee (caller, callee))
2639 free (callee);
2640 return TRUE;
2641 }
2642
2643 /* We're only interested in code sections. Testing SEC_IN_MEMORY excludes
2644 overlay stub sections. */
2645
2646 static bfd_boolean
2647 interesting_section (asection *s)
2648 {
2649 return (s->output_section != bfd_abs_section_ptr
2650 && ((s->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_IN_MEMORY))
2651 == (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2652 && s->size != 0);
2653 }
2654
2655 /* Rummage through the relocs for SEC, looking for function calls.
2656 If CALL_TREE is true, fill in call graph. If CALL_TREE is false,
2657 mark destination symbols on calls as being functions. Also
2658 look at branches, which may be tail calls or go to hot/cold
2659 section part of same function. */
2660
2661 static bfd_boolean
2662 mark_functions_via_relocs (asection *sec,
2663 struct bfd_link_info *info,
2664 int call_tree)
2665 {
2666 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
2667 Elf_Internal_Shdr *symtab_hdr;
2668 void *psyms;
2669 unsigned int priority = 0;
2670 static bfd_boolean warned;
2671
2672 if (!interesting_section (sec)
2673 || sec->reloc_count == 0)
2674 return TRUE;
2675
2676 internal_relocs = _bfd_elf_link_read_relocs (sec->owner, sec, NULL, NULL,
2677 info->keep_memory);
2678 if (internal_relocs == NULL)
2679 return FALSE;
2680
2681 symtab_hdr = &elf_tdata (sec->owner)->symtab_hdr;
2682 psyms = &symtab_hdr->contents;
2683 irela = internal_relocs;
2684 irelaend = irela + sec->reloc_count;
2685 for (; irela < irelaend; irela++)
2686 {
2687 enum elf_spu_reloc_type r_type;
2688 unsigned int r_indx;
2689 asection *sym_sec;
2690 Elf_Internal_Sym *sym;
2691 struct elf_link_hash_entry *h;
2692 bfd_vma val;
2693 bfd_boolean nonbranch, is_call;
2694 struct function_info *caller;
2695 struct call_info *callee;
2696
2697 r_type = ELF32_R_TYPE (irela->r_info);
2698 nonbranch = r_type != R_SPU_REL16 && r_type != R_SPU_ADDR16;
2699
2700 r_indx = ELF32_R_SYM (irela->r_info);
2701 if (!get_sym_h (&h, &sym, &sym_sec, psyms, r_indx, sec->owner))
2702 return FALSE;
2703
2704 if (sym_sec == NULL
2705 || sym_sec->output_section == bfd_abs_section_ptr)
2706 continue;
2707
2708 is_call = FALSE;
2709 if (!nonbranch)
2710 {
2711 unsigned char insn[4];
2712
2713 if (!bfd_get_section_contents (sec->owner, sec, insn,
2714 irela->r_offset, 4))
2715 return FALSE;
2716 if (is_branch (insn))
2717 {
2718 is_call = (insn[0] & 0xfd) == 0x31;
2719 priority = insn[1] & 0x0f;
2720 priority <<= 8;
2721 priority |= insn[2];
2722 priority <<= 8;
2723 priority |= insn[3];
2724 priority >>= 7;
2725 if ((sym_sec->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2726 != (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2727 {
2728 if (!warned)
2729 info->callbacks->einfo
2730 (_("%B(%A+0x%v): call to non-code section"
2731 " %B(%A), analysis incomplete\n"),
2732 sec->owner, sec, irela->r_offset,
2733 sym_sec->owner, sym_sec);
2734 warned = TRUE;
2735 continue;
2736 }
2737 }
2738 else
2739 {
2740 nonbranch = TRUE;
2741 if (is_hint (insn))
2742 continue;
2743 }
2744 }
2745
2746 if (nonbranch)
2747 {
2748 /* For --auto-overlay, count possible stubs we need for
2749 function pointer references. */
2750 unsigned int sym_type;
2751 if (h)
2752 sym_type = h->type;
2753 else
2754 sym_type = ELF_ST_TYPE (sym->st_info);
2755 if (sym_type == STT_FUNC)
2756 {
2757 if (call_tree && spu_hash_table (info)->params->auto_overlay)
2758 spu_hash_table (info)->non_ovly_stub += 1;
2759 /* If the symbol type is STT_FUNC then this must be a
2760 function pointer initialisation. */
2761 continue;
2762 }
2763 /* Ignore data references. */
2764 if ((sym_sec->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2765 != (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2766 continue;
2767 /* Otherwise we probably have a jump table reloc for
2768 a switch statement or some other reference to a
2769 code label. */
2770 }
2771
2772 if (h)
2773 val = h->root.u.def.value;
2774 else
2775 val = sym->st_value;
2776 val += irela->r_addend;
2777
2778 if (!call_tree)
2779 {
2780 struct function_info *fun;
2781
2782 if (irela->r_addend != 0)
2783 {
2784 Elf_Internal_Sym *fake = bfd_zmalloc (sizeof (*fake));
2785 if (fake == NULL)
2786 return FALSE;
2787 fake->st_value = val;
2788 fake->st_shndx
2789 = _bfd_elf_section_from_bfd_section (sym_sec->owner, sym_sec);
2790 sym = fake;
2791 }
2792 if (sym)
2793 fun = maybe_insert_function (sym_sec, sym, FALSE, is_call);
2794 else
2795 fun = maybe_insert_function (sym_sec, h, TRUE, is_call);
2796 if (fun == NULL)
2797 return FALSE;
2798 if (irela->r_addend != 0
2799 && fun->u.sym != sym)
2800 free (sym);
2801 continue;
2802 }
2803
2804 caller = find_function (sec, irela->r_offset, info);
2805 if (caller == NULL)
2806 return FALSE;
2807 callee = bfd_malloc (sizeof *callee);
2808 if (callee == NULL)
2809 return FALSE;
2810
2811 callee->fun = find_function (sym_sec, val, info);
2812 if (callee->fun == NULL)
2813 return FALSE;
2814 callee->is_tail = !is_call;
2815 callee->is_pasted = FALSE;
2816 callee->broken_cycle = FALSE;
2817 callee->priority = priority;
2818 callee->count = nonbranch? 0 : 1;
2819 if (callee->fun->last_caller != sec)
2820 {
2821 callee->fun->last_caller = sec;
2822 callee->fun->call_count += 1;
2823 }
2824 if (!insert_callee (caller, callee))
2825 free (callee);
2826 else if (!is_call
2827 && !callee->fun->is_func
2828 && callee->fun->stack == 0)
2829 {
2830 /* This is either a tail call or a branch from one part of
2831 the function to another, ie. hot/cold section. If the
2832 destination has been called by some other function then
2833 it is a separate function. We also assume that functions
2834 are not split across input files. */
2835 if (sec->owner != sym_sec->owner)
2836 {
2837 callee->fun->start = NULL;
2838 callee->fun->is_func = TRUE;
2839 }
2840 else if (callee->fun->start == NULL)
2841 {
2842 struct function_info *caller_start = caller;
2843 while (caller_start->start)
2844 caller_start = caller_start->start;
2845
2846 if (caller_start != callee->fun)
2847 callee->fun->start = caller_start;
2848 }
2849 else
2850 {
2851 struct function_info *callee_start;
2852 struct function_info *caller_start;
2853 callee_start = callee->fun;
2854 while (callee_start->start)
2855 callee_start = callee_start->start;
2856 caller_start = caller;
2857 while (caller_start->start)
2858 caller_start = caller_start->start;
2859 if (caller_start != callee_start)
2860 {
2861 callee->fun->start = NULL;
2862 callee->fun->is_func = TRUE;
2863 }
2864 }
2865 }
2866 }
2867
2868 return TRUE;
2869 }
2870
2871 /* Handle something like .init or .fini, which has a piece of a function.
2872 These sections are pasted together to form a single function. */
2873
2874 static bfd_boolean
2875 pasted_function (asection *sec)
2876 {
2877 struct bfd_link_order *l;
2878 struct _spu_elf_section_data *sec_data;
2879 struct spu_elf_stack_info *sinfo;
2880 Elf_Internal_Sym *fake;
2881 struct function_info *fun, *fun_start;
2882
2883 fake = bfd_zmalloc (sizeof (*fake));
2884 if (fake == NULL)
2885 return FALSE;
2886 fake->st_value = 0;
2887 fake->st_size = sec->size;
2888 fake->st_shndx
2889 = _bfd_elf_section_from_bfd_section (sec->owner, sec);
2890 fun = maybe_insert_function (sec, fake, FALSE, FALSE);
2891 if (!fun)
2892 return FALSE;
2893
2894 /* Find a function immediately preceding this section. */
2895 fun_start = NULL;
2896 for (l = sec->output_section->map_head.link_order; l != NULL; l = l->next)
2897 {
2898 if (l->u.indirect.section == sec)
2899 {
2900 if (fun_start != NULL)
2901 {
2902 struct call_info *callee = bfd_malloc (sizeof *callee);
2903 if (callee == NULL)
2904 return FALSE;
2905
2906 fun->start = fun_start;
2907 callee->fun = fun;
2908 callee->is_tail = TRUE;
2909 callee->is_pasted = TRUE;
2910 callee->broken_cycle = FALSE;
2911 callee->priority = 0;
2912 callee->count = 1;
2913 if (!insert_callee (fun_start, callee))
2914 free (callee);
2915 return TRUE;
2916 }
2917 break;
2918 }
2919 if (l->type == bfd_indirect_link_order
2920 && (sec_data = spu_elf_section_data (l->u.indirect.section)) != NULL
2921 && (sinfo = sec_data->u.i.stack_info) != NULL
2922 && sinfo->num_fun != 0)
2923 fun_start = &sinfo->fun[sinfo->num_fun - 1];
2924 }
2925
2926 /* Don't return an error if we did not find a function preceding this
2927 section. The section may have incorrect flags. */
2928 return TRUE;
2929 }
2930
2931 /* Map address ranges in code sections to functions. */
2932
2933 static bfd_boolean
2934 discover_functions (struct bfd_link_info *info)
2935 {
2936 bfd *ibfd;
2937 int bfd_idx;
2938 Elf_Internal_Sym ***psym_arr;
2939 asection ***sec_arr;
2940 bfd_boolean gaps = FALSE;
2941
2942 bfd_idx = 0;
2943 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2944 bfd_idx++;
2945
2946 psym_arr = bfd_zmalloc (bfd_idx * sizeof (*psym_arr));
2947 if (psym_arr == NULL)
2948 return FALSE;
2949 sec_arr = bfd_zmalloc (bfd_idx * sizeof (*sec_arr));
2950 if (sec_arr == NULL)
2951 return FALSE;
2952
2953 for (ibfd = info->input_bfds, bfd_idx = 0;
2954 ibfd != NULL;
2955 ibfd = ibfd->link_next, bfd_idx++)
2956 {
2957 extern const bfd_target bfd_elf32_spu_vec;
2958 Elf_Internal_Shdr *symtab_hdr;
2959 asection *sec;
2960 size_t symcount;
2961 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
2962 asection **psecs, **p;
2963
2964 if (ibfd->xvec != &bfd_elf32_spu_vec)
2965 continue;
2966
2967 /* Read all the symbols. */
2968 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2969 symcount = symtab_hdr->sh_size / symtab_hdr->sh_entsize;
2970 if (symcount == 0)
2971 {
2972 if (!gaps)
2973 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2974 if (interesting_section (sec))
2975 {
2976 gaps = TRUE;
2977 break;
2978 }
2979 continue;
2980 }
2981
2982 if (symtab_hdr->contents != NULL)
2983 {
2984 /* Don't use cached symbols since the generic ELF linker
2985 code only reads local symbols, and we need globals too. */
2986 free (symtab_hdr->contents);
2987 symtab_hdr->contents = NULL;
2988 }
2989 syms = bfd_elf_get_elf_syms (ibfd, symtab_hdr, symcount, 0,
2990 NULL, NULL, NULL);
2991 symtab_hdr->contents = (void *) syms;
2992 if (syms == NULL)
2993 return FALSE;
2994
2995 /* Select defined function symbols that are going to be output. */
2996 psyms = bfd_malloc ((symcount + 1) * sizeof (*psyms));
2997 if (psyms == NULL)
2998 return FALSE;
2999 psym_arr[bfd_idx] = psyms;
3000 psecs = bfd_malloc (symcount * sizeof (*psecs));
3001 if (psecs == NULL)
3002 return FALSE;
3003 sec_arr[bfd_idx] = psecs;
3004 for (psy = psyms, p = psecs, sy = syms; sy < syms + symcount; ++p, ++sy)
3005 if (ELF_ST_TYPE (sy->st_info) == STT_NOTYPE
3006 || ELF_ST_TYPE (sy->st_info) == STT_FUNC)
3007 {
3008 asection *s;
3009
3010 *p = s = bfd_section_from_elf_index (ibfd, sy->st_shndx);
3011 if (s != NULL && interesting_section (s))
3012 *psy++ = sy;
3013 }
3014 symcount = psy - psyms;
3015 *psy = NULL;
3016
3017 /* Sort them by section and offset within section. */
3018 sort_syms_syms = syms;
3019 sort_syms_psecs = psecs;
3020 qsort (psyms, symcount, sizeof (*psyms), sort_syms);
3021
3022 /* Now inspect the function symbols. */
3023 for (psy = psyms; psy < psyms + symcount; )
3024 {
3025 asection *s = psecs[*psy - syms];
3026 Elf_Internal_Sym **psy2;
3027
3028 for (psy2 = psy; ++psy2 < psyms + symcount; )
3029 if (psecs[*psy2 - syms] != s)
3030 break;
3031
3032 if (!alloc_stack_info (s, psy2 - psy))
3033 return FALSE;
3034 psy = psy2;
3035 }
3036
3037 /* First install info about properly typed and sized functions.
3038 In an ideal world this will cover all code sections, except
3039 when partitioning functions into hot and cold sections,
3040 and the horrible pasted together .init and .fini functions. */
3041 for (psy = psyms; psy < psyms + symcount; ++psy)
3042 {
3043 sy = *psy;
3044 if (ELF_ST_TYPE (sy->st_info) == STT_FUNC)
3045 {
3046 asection *s = psecs[sy - syms];
3047 if (!maybe_insert_function (s, sy, FALSE, TRUE))
3048 return FALSE;
3049 }
3050 }
3051
3052 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
3053 if (interesting_section (sec))
3054 gaps |= check_function_ranges (sec, info);
3055 }
3056
3057 if (gaps)
3058 {
3059 /* See if we can discover more function symbols by looking at
3060 relocations. */
3061 for (ibfd = info->input_bfds, bfd_idx = 0;
3062 ibfd != NULL;
3063 ibfd = ibfd->link_next, bfd_idx++)
3064 {
3065 asection *sec;
3066
3067 if (psym_arr[bfd_idx] == NULL)
3068 continue;
3069
3070 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3071 if (!mark_functions_via_relocs (sec, info, FALSE))
3072 return FALSE;
3073 }
3074
3075 for (ibfd = info->input_bfds, bfd_idx = 0;
3076 ibfd != NULL;
3077 ibfd = ibfd->link_next, bfd_idx++)
3078 {
3079 Elf_Internal_Shdr *symtab_hdr;
3080 asection *sec;
3081 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
3082 asection **psecs;
3083
3084 if ((psyms = psym_arr[bfd_idx]) == NULL)
3085 continue;
3086
3087 psecs = sec_arr[bfd_idx];
3088
3089 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
3090 syms = (Elf_Internal_Sym *) symtab_hdr->contents;
3091
3092 gaps = FALSE;
3093 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
3094 if (interesting_section (sec))
3095 gaps |= check_function_ranges (sec, info);
3096 if (!gaps)
3097 continue;
3098
3099 /* Finally, install all globals. */
3100 for (psy = psyms; (sy = *psy) != NULL; ++psy)
3101 {
3102 asection *s;
3103
3104 s = psecs[sy - syms];
3105
3106 /* Global syms might be improperly typed functions. */
3107 if (ELF_ST_TYPE (sy->st_info) != STT_FUNC
3108 && ELF_ST_BIND (sy->st_info) == STB_GLOBAL)
3109 {
3110 if (!maybe_insert_function (s, sy, FALSE, FALSE))
3111 return FALSE;
3112 }
3113 }
3114 }
3115
3116 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3117 {
3118 extern const bfd_target bfd_elf32_spu_vec;
3119 asection *sec;
3120
3121 if (ibfd->xvec != &bfd_elf32_spu_vec)
3122 continue;
3123
3124 /* Some of the symbols we've installed as marking the
3125 beginning of functions may have a size of zero. Extend
3126 the range of such functions to the beginning of the
3127 next symbol of interest. */
3128 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3129 if (interesting_section (sec))
3130 {
3131 struct _spu_elf_section_data *sec_data;
3132 struct spu_elf_stack_info *sinfo;
3133
3134 sec_data = spu_elf_section_data (sec);
3135 sinfo = sec_data->u.i.stack_info;
3136 if (sinfo != NULL && sinfo->num_fun != 0)
3137 {
3138 int fun_idx;
3139 bfd_vma hi = sec->size;
3140
3141 for (fun_idx = sinfo->num_fun; --fun_idx >= 0; )
3142 {
3143 sinfo->fun[fun_idx].hi = hi;
3144 hi = sinfo->fun[fun_idx].lo;
3145 }
3146
3147 sinfo->fun[0].lo = 0;
3148 }
3149 /* No symbols in this section. Must be .init or .fini
3150 or something similar. */
3151 else if (!pasted_function (sec))
3152 return FALSE;
3153 }
3154 }
3155 }
3156
3157 for (ibfd = info->input_bfds, bfd_idx = 0;
3158 ibfd != NULL;
3159 ibfd = ibfd->link_next, bfd_idx++)
3160 {
3161 if (psym_arr[bfd_idx] == NULL)
3162 continue;
3163
3164 free (psym_arr[bfd_idx]);
3165 free (sec_arr[bfd_idx]);
3166 }
3167
3168 free (psym_arr);
3169 free (sec_arr);
3170
3171 return TRUE;
3172 }
3173
3174 /* Iterate over all function_info we have collected, calling DOIT on
3175 each node if ROOT_ONLY is false. Only call DOIT on root nodes
3176 if ROOT_ONLY. */
3177
3178 static bfd_boolean
3179 for_each_node (bfd_boolean (*doit) (struct function_info *,
3180 struct bfd_link_info *,
3181 void *),
3182 struct bfd_link_info *info,
3183 void *param,
3184 int root_only)
3185 {
3186 bfd *ibfd;
3187
3188 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3189 {
3190 extern const bfd_target bfd_elf32_spu_vec;
3191 asection *sec;
3192
3193 if (ibfd->xvec != &bfd_elf32_spu_vec)
3194 continue;
3195
3196 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3197 {
3198 struct _spu_elf_section_data *sec_data;
3199 struct spu_elf_stack_info *sinfo;
3200
3201 if ((sec_data = spu_elf_section_data (sec)) != NULL
3202 && (sinfo = sec_data->u.i.stack_info) != NULL)
3203 {
3204 int i;
3205 for (i = 0; i < sinfo->num_fun; ++i)
3206 if (!root_only || !sinfo->fun[i].non_root)
3207 if (!doit (&sinfo->fun[i], info, param))
3208 return FALSE;
3209 }
3210 }
3211 }
3212 return TRUE;
3213 }
3214
3215 /* Transfer call info attached to struct function_info entries for
3216 all of a given function's sections to the first entry. */
3217
3218 static bfd_boolean
3219 transfer_calls (struct function_info *fun,
3220 struct bfd_link_info *info ATTRIBUTE_UNUSED,
3221 void *param ATTRIBUTE_UNUSED)
3222 {
3223 struct function_info *start = fun->start;
3224
3225 if (start != NULL)
3226 {
3227 struct call_info *call, *call_next;
3228
3229 while (start->start != NULL)
3230 start = start->start;
3231 for (call = fun->call_list; call != NULL; call = call_next)
3232 {
3233 call_next = call->next;
3234 if (!insert_callee (start, call))
3235 free (call);
3236 }
3237 fun->call_list = NULL;
3238 }
3239 return TRUE;
3240 }
3241
3242 /* Mark nodes in the call graph that are called by some other node. */
3243
3244 static bfd_boolean
3245 mark_non_root (struct function_info *fun,
3246 struct bfd_link_info *info ATTRIBUTE_UNUSED,
3247 void *param ATTRIBUTE_UNUSED)
3248 {
3249 struct call_info *call;
3250
3251 if (fun->visit1)
3252 return TRUE;
3253 fun->visit1 = TRUE;
3254 for (call = fun->call_list; call; call = call->next)
3255 {
3256 call->fun->non_root = TRUE;
3257 mark_non_root (call->fun, 0, 0);
3258 }
3259 return TRUE;
3260 }
3261
3262 /* Remove cycles from the call graph. Set depth of nodes. */
3263
3264 static bfd_boolean
3265 remove_cycles (struct function_info *fun,
3266 struct bfd_link_info *info,
3267 void *param)
3268 {
3269 struct call_info **callp, *call;
3270 unsigned int depth = *(unsigned int *) param;
3271 unsigned int max_depth = depth;
3272
3273 fun->depth = depth;
3274 fun->visit2 = TRUE;
3275 fun->marking = TRUE;
3276
3277 callp = &fun->call_list;
3278 while ((call = *callp) != NULL)
3279 {
3280 call->max_depth = depth + !call->is_pasted;
3281 if (!call->fun->visit2)
3282 {
3283 if (!remove_cycles (call->fun, info, &call->max_depth))
3284 return FALSE;
3285 if (max_depth < call->max_depth)
3286 max_depth = call->max_depth;
3287 }
3288 else if (call->fun->marking)
3289 {
3290 struct spu_link_hash_table *htab = spu_hash_table (info);
3291
3292 if (!htab->params->auto_overlay
3293 && htab->params->stack_analysis)
3294 {
3295 const char *f1 = func_name (fun);
3296 const char *f2 = func_name (call->fun);
3297
3298 info->callbacks->info (_("Stack analysis will ignore the call "
3299 "from %s to %s\n"),
3300 f1, f2);
3301 }
3302
3303 call->broken_cycle = TRUE;
3304 }
3305 callp = &call->next;
3306 }
3307 fun->marking = FALSE;
3308 *(unsigned int *) param = max_depth;
3309 return TRUE;
3310 }
3311
3312 /* Check that we actually visited all nodes in remove_cycles. If we
3313 didn't, then there is some cycle in the call graph not attached to
3314 any root node. Arbitrarily choose a node in the cycle as a new
3315 root and break the cycle. */
3316
3317 static bfd_boolean
3318 mark_detached_root (struct function_info *fun,
3319 struct bfd_link_info *info,
3320 void *param)
3321 {
3322 if (fun->visit2)
3323 return TRUE;
3324 fun->non_root = FALSE;
3325 *(unsigned int *) param = 0;
3326 return remove_cycles (fun, info, param);
3327 }
3328
3329 /* Populate call_list for each function. */
3330
3331 static bfd_boolean
3332 build_call_tree (struct bfd_link_info *info)
3333 {
3334 bfd *ibfd;
3335 unsigned int depth;
3336
3337 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3338 {
3339 extern const bfd_target bfd_elf32_spu_vec;
3340 asection *sec;
3341
3342 if (ibfd->xvec != &bfd_elf32_spu_vec)
3343 continue;
3344
3345 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3346 if (!mark_functions_via_relocs (sec, info, TRUE))
3347 return FALSE;
3348 }
3349
3350 /* Transfer call info from hot/cold section part of function
3351 to main entry. */
3352 if (!spu_hash_table (info)->params->auto_overlay
3353 && !for_each_node (transfer_calls, info, 0, FALSE))
3354 return FALSE;
3355
3356 /* Find the call graph root(s). */
3357 if (!for_each_node (mark_non_root, info, 0, FALSE))
3358 return FALSE;
3359
3360 /* Remove cycles from the call graph. We start from the root node(s)
3361 so that we break cycles in a reasonable place. */
3362 depth = 0;
3363 if (!for_each_node (remove_cycles, info, &depth, TRUE))
3364 return FALSE;
3365
3366 return for_each_node (mark_detached_root, info, &depth, FALSE);
3367 }
3368
3369 /* qsort predicate to sort calls by priority, max_depth then count. */
3370
3371 static int
3372 sort_calls (const void *a, const void *b)
3373 {
3374 struct call_info *const *c1 = a;
3375 struct call_info *const *c2 = b;
3376 int delta;
3377
3378 delta = (*c2)->priority - (*c1)->priority;
3379 if (delta != 0)
3380 return delta;
3381
3382 delta = (*c2)->max_depth - (*c1)->max_depth;
3383 if (delta != 0)
3384 return delta;
3385
3386 delta = (*c2)->count - (*c1)->count;
3387 if (delta != 0)
3388 return delta;
3389
3390 return (char *) c1 - (char *) c2;
3391 }
3392
3393 struct _mos_param {
3394 unsigned int max_overlay_size;
3395 };
3396
3397 /* Set linker_mark and gc_mark on any sections that we will put in
3398 overlays. These flags are used by the generic ELF linker, but we
3399 won't be continuing on to bfd_elf_final_link so it is OK to use
3400 them. linker_mark is clear before we get here. Set segment_mark
3401 on sections that are part of a pasted function (excluding the last
3402 section).
3403
3404 Set up function rodata section if --overlay-rodata. We don't
3405 currently include merged string constant rodata sections since
3406
3407 Sort the call graph so that the deepest nodes will be visited
3408 first. */
3409
3410 static bfd_boolean
3411 mark_overlay_section (struct function_info *fun,
3412 struct bfd_link_info *info,
3413 void *param)
3414 {
3415 struct call_info *call;
3416 unsigned int count;
3417 struct _mos_param *mos_param = param;
3418 struct spu_link_hash_table *htab = spu_hash_table (info);
3419
3420 if (fun->visit4)
3421 return TRUE;
3422
3423 fun->visit4 = TRUE;
3424 if (!fun->sec->linker_mark
3425 && (htab->params->ovly_flavour != ovly_soft_icache
3426 || htab->params->non_ia_text
3427 || strncmp (fun->sec->name, ".text.ia.", 9) == 0
3428 || strcmp (fun->sec->name, ".init") == 0
3429 || strcmp (fun->sec->name, ".fini") == 0))
3430 {
3431 unsigned int size;
3432
3433 fun->sec->linker_mark = 1;
3434 fun->sec->gc_mark = 1;
3435 fun->sec->segment_mark = 0;
3436 /* Ensure SEC_CODE is set on this text section (it ought to
3437 be!), and SEC_CODE is clear on rodata sections. We use
3438 this flag to differentiate the two overlay section types. */
3439 fun->sec->flags |= SEC_CODE;
3440
3441 size = fun->sec->size;
3442 if (htab->params->auto_overlay & OVERLAY_RODATA)
3443 {
3444 char *name = NULL;
3445
3446 /* Find the rodata section corresponding to this function's
3447 text section. */
3448 if (strcmp (fun->sec->name, ".text") == 0)
3449 {
3450 name = bfd_malloc (sizeof (".rodata"));
3451 if (name == NULL)
3452 return FALSE;
3453 memcpy (name, ".rodata", sizeof (".rodata"));
3454 }
3455 else if (strncmp (fun->sec->name, ".text.", 6) == 0)
3456 {
3457 size_t len = strlen (fun->sec->name);
3458 name = bfd_malloc (len + 3);
3459 if (name == NULL)
3460 return FALSE;
3461 memcpy (name, ".rodata", sizeof (".rodata"));
3462 memcpy (name + 7, fun->sec->name + 5, len - 4);
3463 }
3464 else if (strncmp (fun->sec->name, ".gnu.linkonce.t.", 16) == 0)
3465 {
3466 size_t len = strlen (fun->sec->name) + 1;
3467 name = bfd_malloc (len);
3468 if (name == NULL)
3469 return FALSE;
3470 memcpy (name, fun->sec->name, len);
3471 name[14] = 'r';
3472 }
3473
3474 if (name != NULL)
3475 {
3476 asection *rodata = NULL;
3477 asection *group_sec = elf_section_data (fun->sec)->next_in_group;
3478 if (group_sec == NULL)
3479 rodata = bfd_get_section_by_name (fun->sec->owner, name);
3480 else
3481 while (group_sec != NULL && group_sec != fun->sec)
3482 {
3483 if (strcmp (group_sec->name, name) == 0)
3484 {
3485 rodata = group_sec;
3486 break;
3487 }
3488 group_sec = elf_section_data (group_sec)->next_in_group;
3489 }
3490 fun->rodata = rodata;
3491 if (fun->rodata)
3492 {
3493 size += fun->rodata->size;
3494 if (htab->params->line_size != 0
3495 && size > htab->params->line_size)
3496 {
3497 size -= fun->rodata->size;
3498 fun->rodata = NULL;
3499 }
3500 else
3501 {
3502 fun->rodata->linker_mark = 1;
3503 fun->rodata->gc_mark = 1;
3504 fun->rodata->flags &= ~SEC_CODE;
3505 }
3506 }
3507 free (name);
3508 }
3509 }
3510 if (mos_param->max_overlay_size < size)
3511 mos_param->max_overlay_size = size;
3512 }
3513
3514 for (count = 0, call = fun->call_list; call != NULL; call = call->next)
3515 count += 1;
3516
3517 if (count > 1)
3518 {
3519 struct call_info **calls = bfd_malloc (count * sizeof (*calls));
3520 if (calls == NULL)
3521 return FALSE;
3522
3523 for (count = 0, call = fun->call_list; call != NULL; call = call->next)
3524 calls[count++] = call;
3525
3526 qsort (calls, count, sizeof (*calls), sort_calls);
3527
3528 fun->call_list = NULL;
3529 while (count != 0)
3530 {
3531 --count;
3532 calls[count]->next = fun->call_list;
3533 fun->call_list = calls[count];
3534 }
3535 free (calls);
3536 }
3537
3538 for (call = fun->call_list; call != NULL; call = call->next)
3539 {
3540 if (call->is_pasted)
3541 {
3542 /* There can only be one is_pasted call per function_info. */
3543 BFD_ASSERT (!fun->sec->segment_mark);
3544 fun->sec->segment_mark = 1;
3545 }
3546 if (!call->broken_cycle
3547 && !mark_overlay_section (call->fun, info, param))
3548 return FALSE;
3549 }
3550
3551 /* Don't put entry code into an overlay. The overlay manager needs
3552 a stack! Also, don't mark .ovl.init as an overlay. */
3553 if (fun->lo + fun->sec->output_offset + fun->sec->output_section->vma
3554 == info->output_bfd->start_address
3555 || strncmp (fun->sec->output_section->name, ".ovl.init", 9) == 0)
3556 {
3557 fun->sec->linker_mark = 0;
3558 if (fun->rodata != NULL)
3559 fun->rodata->linker_mark = 0;
3560 }
3561 return TRUE;
3562 }
3563
3564 /* If non-zero then unmark functions called from those within sections
3565 that we need to unmark. Unfortunately this isn't reliable since the
3566 call graph cannot know the destination of function pointer calls. */
3567 #define RECURSE_UNMARK 0
3568
3569 struct _uos_param {
3570 asection *exclude_input_section;
3571 asection *exclude_output_section;
3572 unsigned long clearing;
3573 };
3574
3575 /* Undo some of mark_overlay_section's work. */
3576
3577 static bfd_boolean
3578 unmark_overlay_section (struct function_info *fun,
3579 struct bfd_link_info *info,
3580 void *param)
3581 {
3582 struct call_info *call;
3583 struct _uos_param *uos_param = param;
3584 unsigned int excluded = 0;
3585
3586 if (fun->visit5)
3587 return TRUE;
3588
3589 fun->visit5 = TRUE;
3590
3591 excluded = 0;
3592 if (fun->sec == uos_param->exclude_input_section
3593 || fun->sec->output_section == uos_param->exclude_output_section)
3594 excluded = 1;
3595
3596 if (RECURSE_UNMARK)
3597 uos_param->clearing += excluded;
3598
3599 if (RECURSE_UNMARK ? uos_param->clearing : excluded)
3600 {
3601 fun->sec->linker_mark = 0;
3602 if (fun->rodata)
3603 fun->rodata->linker_mark = 0;
3604 }
3605
3606 for (call = fun->call_list; call != NULL; call = call->next)
3607 if (!call->broken_cycle
3608 && !unmark_overlay_section (call->fun, info, param))
3609 return FALSE;
3610
3611 if (RECURSE_UNMARK)
3612 uos_param->clearing -= excluded;
3613 return TRUE;
3614 }
3615
3616 struct _cl_param {
3617 unsigned int lib_size;
3618 asection **lib_sections;
3619 };
3620
3621 /* Add sections we have marked as belonging to overlays to an array
3622 for consideration as non-overlay sections. The array consist of
3623 pairs of sections, (text,rodata), for functions in the call graph. */
3624
3625 static bfd_boolean
3626 collect_lib_sections (struct function_info *fun,
3627 struct bfd_link_info *info,
3628 void *param)
3629 {
3630 struct _cl_param *lib_param = param;
3631 struct call_info *call;
3632 unsigned int size;
3633
3634 if (fun->visit6)
3635 return TRUE;
3636
3637 fun->visit6 = TRUE;
3638 if (!fun->sec->linker_mark || !fun->sec->gc_mark || fun->sec->segment_mark)
3639 return TRUE;
3640
3641 size = fun->sec->size;
3642 if (fun->rodata)
3643 size += fun->rodata->size;
3644
3645 if (size <= lib_param->lib_size)
3646 {
3647 *lib_param->lib_sections++ = fun->sec;
3648 fun->sec->gc_mark = 0;
3649 if (fun->rodata && fun->rodata->linker_mark && fun->rodata->gc_mark)
3650 {
3651 *lib_param->lib_sections++ = fun->rodata;
3652 fun->rodata->gc_mark = 0;
3653 }
3654 else
3655 *lib_param->lib_sections++ = NULL;
3656 }
3657
3658 for (call = fun->call_list; call != NULL; call = call->next)
3659 if (!call->broken_cycle)
3660 collect_lib_sections (call->fun, info, param);
3661
3662 return TRUE;
3663 }
3664
3665 /* qsort predicate to sort sections by call count. */
3666
3667 static int
3668 sort_lib (const void *a, const void *b)
3669 {
3670 asection *const *s1 = a;
3671 asection *const *s2 = b;
3672 struct _spu_elf_section_data *sec_data;
3673 struct spu_elf_stack_info *sinfo;
3674 int delta;
3675
3676 delta = 0;
3677 if ((sec_data = spu_elf_section_data (*s1)) != NULL
3678 && (sinfo = sec_data->u.i.stack_info) != NULL)
3679 {
3680 int i;
3681 for (i = 0; i < sinfo->num_fun; ++i)
3682 delta -= sinfo->fun[i].call_count;
3683 }
3684
3685 if ((sec_data = spu_elf_section_data (*s2)) != NULL
3686 && (sinfo = sec_data->u.i.stack_info) != NULL)
3687 {
3688 int i;
3689 for (i = 0; i < sinfo->num_fun; ++i)
3690 delta += sinfo->fun[i].call_count;
3691 }
3692
3693 if (delta != 0)
3694 return delta;
3695
3696 return s1 - s2;
3697 }
3698
3699 /* Remove some sections from those marked to be in overlays. Choose
3700 those that are called from many places, likely library functions. */
3701
3702 static unsigned int
3703 auto_ovl_lib_functions (struct bfd_link_info *info, unsigned int lib_size)
3704 {
3705 bfd *ibfd;
3706 asection **lib_sections;
3707 unsigned int i, lib_count;
3708 struct _cl_param collect_lib_param;
3709 struct function_info dummy_caller;
3710 struct spu_link_hash_table *htab;
3711
3712 memset (&dummy_caller, 0, sizeof (dummy_caller));
3713 lib_count = 0;
3714 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3715 {
3716 extern const bfd_target bfd_elf32_spu_vec;
3717 asection *sec;
3718
3719 if (ibfd->xvec != &bfd_elf32_spu_vec)
3720 continue;
3721
3722 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3723 if (sec->linker_mark
3724 && sec->size < lib_size
3725 && (sec->flags & SEC_CODE) != 0)
3726 lib_count += 1;
3727 }
3728 lib_sections = bfd_malloc (lib_count * 2 * sizeof (*lib_sections));
3729 if (lib_sections == NULL)
3730 return (unsigned int) -1;
3731 collect_lib_param.lib_size = lib_size;
3732 collect_lib_param.lib_sections = lib_sections;
3733 if (!for_each_node (collect_lib_sections, info, &collect_lib_param,
3734 TRUE))
3735 return (unsigned int) -1;
3736 lib_count = (collect_lib_param.lib_sections - lib_sections) / 2;
3737
3738 /* Sort sections so that those with the most calls are first. */
3739 if (lib_count > 1)
3740 qsort (lib_sections, lib_count, 2 * sizeof (*lib_sections), sort_lib);
3741
3742 htab = spu_hash_table (info);
3743 for (i = 0; i < lib_count; i++)
3744 {
3745 unsigned int tmp, stub_size;
3746 asection *sec;
3747 struct _spu_elf_section_data *sec_data;
3748 struct spu_elf_stack_info *sinfo;
3749
3750 sec = lib_sections[2 * i];
3751 /* If this section is OK, its size must be less than lib_size. */
3752 tmp = sec->size;
3753 /* If it has a rodata section, then add that too. */
3754 if (lib_sections[2 * i + 1])
3755 tmp += lib_sections[2 * i + 1]->size;
3756 /* Add any new overlay call stubs needed by the section. */
3757 stub_size = 0;
3758 if (tmp < lib_size
3759 && (sec_data = spu_elf_section_data (sec)) != NULL
3760 && (sinfo = sec_data->u.i.stack_info) != NULL)
3761 {
3762 int k;
3763 struct call_info *call;
3764
3765 for (k = 0; k < sinfo->num_fun; ++k)
3766 for (call = sinfo->fun[k].call_list; call; call = call->next)
3767 if (call->fun->sec->linker_mark)
3768 {
3769 struct call_info *p;
3770 for (p = dummy_caller.call_list; p; p = p->next)
3771 if (p->fun == call->fun)
3772 break;
3773 if (!p)
3774 stub_size += ovl_stub_size (htab->params);
3775 }
3776 }
3777 if (tmp + stub_size < lib_size)
3778 {
3779 struct call_info **pp, *p;
3780
3781 /* This section fits. Mark it as non-overlay. */
3782 lib_sections[2 * i]->linker_mark = 0;
3783 if (lib_sections[2 * i + 1])
3784 lib_sections[2 * i + 1]->linker_mark = 0;
3785 lib_size -= tmp + stub_size;
3786 /* Call stubs to the section we just added are no longer
3787 needed. */
3788 pp = &dummy_caller.call_list;
3789 while ((p = *pp) != NULL)
3790 if (!p->fun->sec->linker_mark)
3791 {
3792 lib_size += ovl_stub_size (htab->params);
3793 *pp = p->next;
3794 free (p);
3795 }
3796 else
3797 pp = &p->next;
3798 /* Add new call stubs to dummy_caller. */
3799 if ((sec_data = spu_elf_section_data (sec)) != NULL
3800 && (sinfo = sec_data->u.i.stack_info) != NULL)
3801 {
3802 int k;
3803 struct call_info *call;
3804
3805 for (k = 0; k < sinfo->num_fun; ++k)
3806 for (call = sinfo->fun[k].call_list;
3807 call;
3808 call = call->next)
3809 if (call->fun->sec->linker_mark)
3810 {
3811 struct call_info *callee;
3812 callee = bfd_malloc (sizeof (*callee));
3813 if (callee == NULL)
3814 return (unsigned int) -1;
3815 *callee = *call;
3816 if (!insert_callee (&dummy_caller, callee))
3817 free (callee);
3818 }
3819 }
3820 }
3821 }
3822 while (dummy_caller.call_list != NULL)
3823 {
3824 struct call_info *call = dummy_caller.call_list;
3825 dummy_caller.call_list = call->next;
3826 free (call);
3827 }
3828 for (i = 0; i < 2 * lib_count; i++)
3829 if (lib_sections[i])
3830 lib_sections[i]->gc_mark = 1;
3831 free (lib_sections);
3832 return lib_size;
3833 }
3834
3835 /* Build an array of overlay sections. The deepest node's section is
3836 added first, then its parent node's section, then everything called
3837 from the parent section. The idea being to group sections to
3838 minimise calls between different overlays. */
3839
3840 static bfd_boolean
3841 collect_overlays (struct function_info *fun,
3842 struct bfd_link_info *info,
3843 void *param)
3844 {
3845 struct call_info *call;
3846 bfd_boolean added_fun;
3847 asection ***ovly_sections = param;
3848
3849 if (fun->visit7)
3850 return TRUE;
3851
3852 fun->visit7 = TRUE;
3853 for (call = fun->call_list; call != NULL; call = call->next)
3854 if (!call->is_pasted && !call->broken_cycle)
3855 {
3856 if (!collect_overlays (call->fun, info, ovly_sections))
3857 return FALSE;
3858 break;
3859 }
3860
3861 added_fun = FALSE;
3862 if (fun->sec->linker_mark && fun->sec->gc_mark)
3863 {
3864 fun->sec->gc_mark = 0;
3865 *(*ovly_sections)++ = fun->sec;
3866 if (fun->rodata && fun->rodata->linker_mark && fun->rodata->gc_mark)
3867 {
3868 fun->rodata->gc_mark = 0;
3869 *(*ovly_sections)++ = fun->rodata;
3870 }
3871 else
3872 *(*ovly_sections)++ = NULL;
3873 added_fun = TRUE;
3874
3875 /* Pasted sections must stay with the first section. We don't
3876 put pasted sections in the array, just the first section.
3877 Mark subsequent sections as already considered. */
3878 if (fun->sec->segment_mark)
3879 {
3880 struct function_info *call_fun = fun;
3881 do
3882 {
3883 for (call = call_fun->call_list; call != NULL; call = call->next)
3884 if (call->is_pasted)
3885 {
3886 call_fun = call->fun;
3887 call_fun->sec->gc_mark = 0;
3888 if (call_fun->rodata)
3889 call_fun->rodata->gc_mark = 0;
3890 break;
3891 }
3892 if (call == NULL)
3893 abort ();
3894 }
3895 while (call_fun->sec->segment_mark);
3896 }
3897 }
3898
3899 for (call = fun->call_list; call != NULL; call = call->next)
3900 if (!call->broken_cycle
3901 && !collect_overlays (call->fun, info, ovly_sections))
3902 return FALSE;
3903
3904 if (added_fun)
3905 {
3906 struct _spu_elf_section_data *sec_data;
3907 struct spu_elf_stack_info *sinfo;
3908
3909 if ((sec_data = spu_elf_section_data (fun->sec)) != NULL
3910 && (sinfo = sec_data->u.i.stack_info) != NULL)
3911 {
3912 int i;
3913 for (i = 0; i < sinfo->num_fun; ++i)
3914 if (!collect_overlays (&sinfo->fun[i], info, ovly_sections))
3915 return FALSE;
3916 }
3917 }
3918
3919 return TRUE;
3920 }
3921
3922 struct _sum_stack_param {
3923 size_t cum_stack;
3924 size_t overall_stack;
3925 bfd_boolean emit_stack_syms;
3926 };
3927
3928 /* Descend the call graph for FUN, accumulating total stack required. */
3929
3930 static bfd_boolean
3931 sum_stack (struct function_info *fun,
3932 struct bfd_link_info *info,
3933 void *param)
3934 {
3935 struct call_info *call;
3936 struct function_info *max;
3937 size_t stack, cum_stack;
3938 const char *f1;
3939 bfd_boolean has_call;
3940 struct _sum_stack_param *sum_stack_param = param;
3941 struct spu_link_hash_table *htab;
3942
3943 cum_stack = fun->stack;
3944 sum_stack_param->cum_stack = cum_stack;
3945 if (fun->visit3)
3946 return TRUE;
3947
3948 has_call = FALSE;
3949 max = NULL;
3950 for (call = fun->call_list; call; call = call->next)
3951 {
3952 if (call->broken_cycle)
3953 continue;
3954 if (!call->is_pasted)
3955 has_call = TRUE;
3956 if (!sum_stack (call->fun, info, sum_stack_param))
3957 return FALSE;
3958 stack = sum_stack_param->cum_stack;
3959 /* Include caller stack for normal calls, don't do so for
3960 tail calls. fun->stack here is local stack usage for
3961 this function. */
3962 if (!call->is_tail || call->is_pasted || call->fun->start != NULL)
3963 stack += fun->stack;
3964 if (cum_stack < stack)
3965 {
3966 cum_stack = stack;
3967 max = call->fun;
3968 }
3969 }
3970
3971 sum_stack_param->cum_stack = cum_stack;
3972 stack = fun->stack;
3973 /* Now fun->stack holds cumulative stack. */
3974 fun->stack = cum_stack;
3975 fun->visit3 = TRUE;
3976
3977 if (!fun->non_root
3978 && sum_stack_param->overall_stack < cum_stack)
3979 sum_stack_param->overall_stack = cum_stack;
3980
3981 htab = spu_hash_table (info);
3982 if (htab->params->auto_overlay)
3983 return TRUE;
3984
3985 f1 = func_name (fun);
3986 if (htab->params->stack_analysis)
3987 {
3988 if (!fun->non_root)
3989 info->callbacks->info (_(" %s: 0x%v\n"), f1, (bfd_vma) cum_stack);
3990 info->callbacks->minfo (_("%s: 0x%v 0x%v\n"),
3991 f1, (bfd_vma) stack, (bfd_vma) cum_stack);
3992
3993 if (has_call)
3994 {
3995 info->callbacks->minfo (_(" calls:\n"));
3996 for (call = fun->call_list; call; call = call->next)
3997 if (!call->is_pasted && !call->broken_cycle)
3998 {
3999 const char *f2 = func_name (call->fun);
4000 const char *ann1 = call->fun == max ? "*" : " ";
4001 const char *ann2 = call->is_tail ? "t" : " ";
4002
4003 info->callbacks->minfo (_(" %s%s %s\n"), ann1, ann2, f2);
4004 }
4005 }
4006 }
4007
4008 if (sum_stack_param->emit_stack_syms)
4009 {
4010 char *name = bfd_malloc (18 + strlen (f1));
4011 struct elf_link_hash_entry *h;
4012
4013 if (name == NULL)
4014 return FALSE;
4015
4016 if (fun->global || ELF_ST_BIND (fun->u.sym->st_info) == STB_GLOBAL)
4017 sprintf (name, "__stack_%s", f1);
4018 else
4019 sprintf (name, "__stack_%x_%s", fun->sec->id & 0xffffffff, f1);
4020
4021 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
4022 free (name);
4023 if (h != NULL
4024 && (h->root.type == bfd_link_hash_new
4025 || h->root.type == bfd_link_hash_undefined
4026 || h->root.type == bfd_link_hash_undefweak))
4027 {
4028 h->root.type = bfd_link_hash_defined;
4029 h->root.u.def.section = bfd_abs_section_ptr;
4030 h->root.u.def.value = cum_stack;
4031 h->size = 0;
4032 h->type = 0;
4033 h->ref_regular = 1;
4034 h->def_regular = 1;
4035 h->ref_regular_nonweak = 1;
4036 h->forced_local = 1;
4037 h->non_elf = 0;
4038 }
4039 }
4040
4041 return TRUE;
4042 }
4043
4044 /* SEC is part of a pasted function. Return the call_info for the
4045 next section of this function. */
4046
4047 static struct call_info *
4048 find_pasted_call (asection *sec)
4049 {
4050 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
4051 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
4052 struct call_info *call;
4053 int k;
4054
4055 for (k = 0; k < sinfo->num_fun; ++k)
4056 for (call = sinfo->fun[k].call_list; call != NULL; call = call->next)
4057 if (call->is_pasted)
4058 return call;
4059 abort ();
4060 return 0;
4061 }
4062
4063 /* qsort predicate to sort bfds by file name. */
4064
4065 static int
4066 sort_bfds (const void *a, const void *b)
4067 {
4068 bfd *const *abfd1 = a;
4069 bfd *const *abfd2 = b;
4070
4071 return filename_cmp ((*abfd1)->filename, (*abfd2)->filename);
4072 }
4073
4074 static unsigned int
4075 print_one_overlay_section (FILE *script,
4076 unsigned int base,
4077 unsigned int count,
4078 unsigned int ovlynum,
4079 unsigned int *ovly_map,
4080 asection **ovly_sections,
4081 struct bfd_link_info *info)
4082 {
4083 unsigned int j;
4084
4085 for (j = base; j < count && ovly_map[j] == ovlynum; j++)
4086 {
4087 asection *sec = ovly_sections[2 * j];
4088
4089 if (fprintf (script, " %s%c%s (%s)\n",
4090 (sec->owner->my_archive != NULL
4091 ? sec->owner->my_archive->filename : ""),
4092 info->path_separator,
4093 sec->owner->filename,
4094 sec->name) <= 0)
4095 return -1;
4096 if (sec->segment_mark)
4097 {
4098 struct call_info *call = find_pasted_call (sec);
4099 while (call != NULL)
4100 {
4101 struct function_info *call_fun = call->fun;
4102 sec = call_fun->sec;
4103 if (fprintf (script, " %s%c%s (%s)\n",
4104 (sec->owner->my_archive != NULL
4105 ? sec->owner->my_archive->filename : ""),
4106 info->path_separator,
4107 sec->owner->filename,
4108 sec->name) <= 0)
4109 return -1;
4110 for (call = call_fun->call_list; call; call = call->next)
4111 if (call->is_pasted)
4112 break;
4113 }
4114 }
4115 }
4116
4117 for (j = base; j < count && ovly_map[j] == ovlynum; j++)
4118 {
4119 asection *sec = ovly_sections[2 * j + 1];
4120 if (sec != NULL
4121 && fprintf (script, " %s%c%s (%s)\n",
4122 (sec->owner->my_archive != NULL
4123 ? sec->owner->my_archive->filename : ""),
4124 info->path_separator,
4125 sec->owner->filename,
4126 sec->name) <= 0)
4127 return -1;
4128
4129 sec = ovly_sections[2 * j];
4130 if (sec->segment_mark)
4131 {
4132 struct call_info *call = find_pasted_call (sec);
4133 while (call != NULL)
4134 {
4135 struct function_info *call_fun = call->fun;
4136 sec = call_fun->rodata;
4137 if (sec != NULL
4138 && fprintf (script, " %s%c%s (%s)\n",
4139 (sec->owner->my_archive != NULL
4140 ? sec->owner->my_archive->filename : ""),
4141 info->path_separator,
4142 sec->owner->filename,
4143 sec->name) <= 0)
4144 return -1;
4145 for (call = call_fun->call_list; call; call = call->next)
4146 if (call->is_pasted)
4147 break;
4148 }
4149 }
4150 }
4151
4152 return j;
4153 }
4154
4155 /* Handle --auto-overlay. */
4156
4157 static void
4158 spu_elf_auto_overlay (struct bfd_link_info *info)
4159 {
4160 bfd *ibfd;
4161 bfd **bfd_arr;
4162 struct elf_segment_map *m;
4163 unsigned int fixed_size, lo, hi;
4164 unsigned int reserved;
4165 struct spu_link_hash_table *htab;
4166 unsigned int base, i, count, bfd_count;
4167 unsigned int region, ovlynum;
4168 asection **ovly_sections, **ovly_p;
4169 unsigned int *ovly_map;
4170 FILE *script;
4171 unsigned int total_overlay_size, overlay_size;
4172 const char *ovly_mgr_entry;
4173 struct elf_link_hash_entry *h;
4174 struct _mos_param mos_param;
4175 struct _uos_param uos_param;
4176 struct function_info dummy_caller;
4177
4178 /* Find the extents of our loadable image. */
4179 lo = (unsigned int) -1;
4180 hi = 0;
4181 for (m = elf_tdata (info->output_bfd)->segment_map; m != NULL; m = m->next)
4182 if (m->p_type == PT_LOAD)
4183 for (i = 0; i < m->count; i++)
4184 if (m->sections[i]->size != 0)
4185 {
4186 if (m->sections[i]->vma < lo)
4187 lo = m->sections[i]->vma;
4188 if (m->sections[i]->vma + m->sections[i]->size - 1 > hi)
4189 hi = m->sections[i]->vma + m->sections[i]->size - 1;
4190 }
4191 fixed_size = hi + 1 - lo;
4192
4193 if (!discover_functions (info))
4194 goto err_exit;
4195
4196 if (!build_call_tree (info))
4197 goto err_exit;
4198
4199 htab = spu_hash_table (info);
4200 reserved = htab->params->auto_overlay_reserved;
4201 if (reserved == 0)
4202 {
4203 struct _sum_stack_param sum_stack_param;
4204
4205 sum_stack_param.emit_stack_syms = 0;
4206 sum_stack_param.overall_stack = 0;
4207 if (!for_each_node (sum_stack, info, &sum_stack_param, TRUE))
4208 goto err_exit;
4209 reserved = (sum_stack_param.overall_stack
4210 + htab->params->extra_stack_space);
4211 }
4212
4213 /* No need for overlays if everything already fits. */
4214 if (fixed_size + reserved <= htab->local_store
4215 && htab->params->ovly_flavour != ovly_soft_icache)
4216 {
4217 htab->params->auto_overlay = 0;
4218 return;
4219 }
4220
4221 uos_param.exclude_input_section = 0;
4222 uos_param.exclude_output_section
4223 = bfd_get_section_by_name (info->output_bfd, ".interrupt");
4224
4225 ovly_mgr_entry = "__ovly_load";
4226 if (htab->params->ovly_flavour == ovly_soft_icache)
4227 ovly_mgr_entry = "__icache_br_handler";
4228 h = elf_link_hash_lookup (&htab->elf, ovly_mgr_entry,
4229 FALSE, FALSE, FALSE);
4230 if (h != NULL
4231 && (h->root.type == bfd_link_hash_defined
4232 || h->root.type == bfd_link_hash_defweak)
4233 && h->def_regular)
4234 {
4235 /* We have a user supplied overlay manager. */
4236 uos_param.exclude_input_section = h->root.u.def.section;
4237 }
4238 else
4239 {
4240 /* If no user overlay manager, spu_elf_load_ovl_mgr will add our
4241 builtin version to .text, and will adjust .text size. */
4242 fixed_size += (*htab->params->spu_elf_load_ovl_mgr) ();
4243 }
4244
4245 /* Mark overlay sections, and find max overlay section size. */
4246 mos_param.max_overlay_size = 0;
4247 if (!for_each_node (mark_overlay_section, info, &mos_param, TRUE))
4248 goto err_exit;
4249
4250 /* We can't put the overlay manager or interrupt routines in
4251 overlays. */
4252 uos_param.clearing = 0;
4253 if ((uos_param.exclude_input_section
4254 || uos_param.exclude_output_section)
4255 && !for_each_node (unmark_overlay_section, info, &uos_param, TRUE))
4256 goto err_exit;
4257
4258 bfd_count = 0;
4259 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
4260 ++bfd_count;
4261 bfd_arr = bfd_malloc (bfd_count * sizeof (*bfd_arr));
4262 if (bfd_arr == NULL)
4263 goto err_exit;
4264
4265 /* Count overlay sections, and subtract their sizes from "fixed_size". */
4266 count = 0;
4267 bfd_count = 0;
4268 total_overlay_size = 0;
4269 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
4270 {
4271 extern const bfd_target bfd_elf32_spu_vec;
4272 asection *sec;
4273 unsigned int old_count;
4274
4275 if (ibfd->xvec != &bfd_elf32_spu_vec)
4276 continue;
4277
4278 old_count = count;
4279 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
4280 if (sec->linker_mark)
4281 {
4282 if ((sec->flags & SEC_CODE) != 0)
4283 count += 1;
4284 fixed_size -= sec->size;
4285 total_overlay_size += sec->size;
4286 }
4287 else if ((sec->flags & (SEC_ALLOC | SEC_LOAD)) == (SEC_ALLOC | SEC_LOAD)
4288 && sec->output_section->owner == info->output_bfd
4289 && strncmp (sec->output_section->name, ".ovl.init", 9) == 0)
4290 fixed_size -= sec->size;
4291 if (count != old_count)
4292 bfd_arr[bfd_count++] = ibfd;
4293 }
4294
4295 /* Since the overlay link script selects sections by file name and
4296 section name, ensure that file names are unique. */
4297 if (bfd_count > 1)
4298 {
4299 bfd_boolean ok = TRUE;
4300
4301 qsort (bfd_arr, bfd_count, sizeof (*bfd_arr), sort_bfds);
4302 for (i = 1; i < bfd_count; ++i)
4303 if (filename_cmp (bfd_arr[i - 1]->filename, bfd_arr[i]->filename) == 0)
4304 {
4305 if (bfd_arr[i - 1]->my_archive == bfd_arr[i]->my_archive)
4306 {
4307 if (bfd_arr[i - 1]->my_archive && bfd_arr[i]->my_archive)
4308 info->callbacks->einfo (_("%s duplicated in %s\n"),
4309 bfd_arr[i]->filename,
4310 bfd_arr[i]->my_archive->filename);
4311 else
4312 info->callbacks->einfo (_("%s duplicated\n"),
4313 bfd_arr[i]->filename);
4314 ok = FALSE;
4315 }
4316 }
4317 if (!ok)
4318 {
4319 info->callbacks->einfo (_("sorry, no support for duplicate "
4320 "object files in auto-overlay script\n"));
4321 bfd_set_error (bfd_error_bad_value);
4322 goto err_exit;
4323 }
4324 }
4325 free (bfd_arr);
4326
4327 fixed_size += reserved;
4328 fixed_size += htab->non_ovly_stub * ovl_stub_size (htab->params);
4329 if (fixed_size + mos_param.max_overlay_size <= htab->local_store)
4330 {
4331 if (htab->params->ovly_flavour == ovly_soft_icache)
4332 {
4333 /* Stubs in the non-icache area are bigger. */
4334 fixed_size += htab->non_ovly_stub * 16;
4335 /* Space for icache manager tables.
4336 a) Tag array, one quadword per cache line.
4337 - word 0: ia address of present line, init to zero. */
4338 fixed_size += 16 << htab->num_lines_log2;
4339 /* b) Rewrite "to" list, one quadword per cache line. */
4340 fixed_size += 16 << htab->num_lines_log2;
4341 /* c) Rewrite "from" list, one byte per outgoing branch (rounded up
4342 to a power-of-two number of full quadwords) per cache line. */
4343 fixed_size += 16 << (htab->fromelem_size_log2
4344 + htab->num_lines_log2);
4345 /* d) Pointer to __ea backing store (toe), 1 quadword. */
4346 fixed_size += 16;
4347 }
4348 else
4349 {
4350 /* Guess number of overlays. Assuming overlay buffer is on
4351 average only half full should be conservative. */
4352 ovlynum = (total_overlay_size * 2 * htab->params->num_lines
4353 / (htab->local_store - fixed_size));
4354 /* Space for _ovly_table[], _ovly_buf_table[] and toe. */
4355 fixed_size += ovlynum * 16 + 16 + 4 + 16;
4356 }
4357 }
4358
4359 if (fixed_size + mos_param.max_overlay_size > htab->local_store)
4360 info->callbacks->einfo (_("non-overlay size of 0x%v plus maximum overlay "
4361 "size of 0x%v exceeds local store\n"),
4362 (bfd_vma) fixed_size,
4363 (bfd_vma) mos_param.max_overlay_size);
4364
4365 /* Now see if we should put some functions in the non-overlay area. */
4366 else if (fixed_size < htab->params->auto_overlay_fixed)
4367 {
4368 unsigned int max_fixed, lib_size;
4369
4370 max_fixed = htab->local_store - mos_param.max_overlay_size;
4371 if (max_fixed > htab->params->auto_overlay_fixed)
4372 max_fixed = htab->params->auto_overlay_fixed;
4373 lib_size = max_fixed - fixed_size;
4374 lib_size = auto_ovl_lib_functions (info, lib_size);
4375 if (lib_size == (unsigned int) -1)
4376 goto err_exit;
4377 fixed_size = max_fixed - lib_size;
4378 }
4379
4380 /* Build an array of sections, suitably sorted to place into
4381 overlays. */
4382 ovly_sections = bfd_malloc (2 * count * sizeof (*ovly_sections));
4383 if (ovly_sections == NULL)
4384 goto err_exit;
4385 ovly_p = ovly_sections;
4386 if (!for_each_node (collect_overlays, info, &ovly_p, TRUE))
4387 goto err_exit;
4388 count = (size_t) (ovly_p - ovly_sections) / 2;
4389 ovly_map = bfd_malloc (count * sizeof (*ovly_map));
4390 if (ovly_map == NULL)
4391 goto err_exit;
4392
4393 memset (&dummy_caller, 0, sizeof (dummy_caller));
4394 overlay_size = (htab->local_store - fixed_size) / htab->params->num_lines;
4395 if (htab->params->line_size != 0)
4396 overlay_size = htab->params->line_size;
4397 base = 0;
4398 ovlynum = 0;
4399 while (base < count)
4400 {
4401 unsigned int size = 0, rosize = 0, roalign = 0;
4402
4403 for (i = base; i < count; i++)
4404 {
4405 asection *sec, *rosec;
4406 unsigned int tmp, rotmp;
4407 unsigned int num_stubs;
4408 struct call_info *call, *pasty;
4409 struct _spu_elf_section_data *sec_data;
4410 struct spu_elf_stack_info *sinfo;
4411 unsigned int k;
4412
4413 /* See whether we can add this section to the current
4414 overlay without overflowing our overlay buffer. */
4415 sec = ovly_sections[2 * i];
4416 tmp = align_power (size, sec->alignment_power) + sec->size;
4417 rotmp = rosize;
4418 rosec = ovly_sections[2 * i + 1];
4419 if (rosec != NULL)
4420 {
4421 rotmp = align_power (rotmp, rosec->alignment_power) + rosec->size;
4422 if (roalign < rosec->alignment_power)
4423 roalign = rosec->alignment_power;
4424 }
4425 if (align_power (tmp, roalign) + rotmp > overlay_size)
4426 break;
4427 if (sec->segment_mark)
4428 {
4429 /* Pasted sections must stay together, so add their
4430 sizes too. */
4431 pasty = find_pasted_call (sec);
4432 while (pasty != NULL)
4433 {
4434 struct function_info *call_fun = pasty->fun;
4435 tmp = (align_power (tmp, call_fun->sec->alignment_power)
4436 + call_fun->sec->size);
4437 if (call_fun->rodata)
4438 {
4439 rotmp = (align_power (rotmp,
4440 call_fun->rodata->alignment_power)
4441 + call_fun->rodata->size);
4442 if (roalign < rosec->alignment_power)
4443 roalign = rosec->alignment_power;
4444 }
4445 for (pasty = call_fun->call_list; pasty; pasty = pasty->next)
4446 if (pasty->is_pasted)
4447 break;
4448 }
4449 }
4450 if (align_power (tmp, roalign) + rotmp > overlay_size)
4451 break;
4452
4453 /* If we add this section, we might need new overlay call
4454 stubs. Add any overlay section calls to dummy_call. */
4455 pasty = NULL;
4456 sec_data = spu_elf_section_data (sec);
4457 sinfo = sec_data->u.i.stack_info;
4458 for (k = 0; k < (unsigned) sinfo->num_fun; ++k)
4459 for (call = sinfo->fun[k].call_list; call; call = call->next)
4460 if (call->is_pasted)
4461 {
4462 BFD_ASSERT (pasty == NULL);
4463 pasty = call;
4464 }
4465 else if (call->fun->sec->linker_mark)
4466 {
4467 if (!copy_callee (&dummy_caller, call))
4468 goto err_exit;
4469 }
4470 while (pasty != NULL)
4471 {
4472 struct function_info *call_fun = pasty->fun;
4473 pasty = NULL;
4474 for (call = call_fun->call_list; call; call = call->next)
4475 if (call->is_pasted)
4476 {
4477 BFD_ASSERT (pasty == NULL);
4478 pasty = call;
4479 }
4480 else if (!copy_callee (&dummy_caller, call))
4481 goto err_exit;
4482 }
4483
4484 /* Calculate call stub size. */
4485 num_stubs = 0;
4486 for (call = dummy_caller.call_list; call; call = call->next)
4487 {
4488 unsigned int stub_delta = 1;
4489
4490 if (htab->params->ovly_flavour == ovly_soft_icache)
4491 stub_delta = call->count;
4492 num_stubs += stub_delta;
4493
4494 /* If the call is within this overlay, we won't need a
4495 stub. */
4496 for (k = base; k < i + 1; k++)
4497 if (call->fun->sec == ovly_sections[2 * k])
4498 {
4499 num_stubs -= stub_delta;
4500 break;
4501 }
4502 }
4503 if (htab->params->ovly_flavour == ovly_soft_icache
4504 && num_stubs > htab->params->max_branch)
4505 break;
4506 if (align_power (tmp, roalign) + rotmp
4507 + num_stubs * ovl_stub_size (htab->params) > overlay_size)
4508 break;
4509 size = tmp;
4510 rosize = rotmp;
4511 }
4512
4513 if (i == base)
4514 {
4515 info->callbacks->einfo (_("%B:%A%s exceeds overlay size\n"),
4516 ovly_sections[2 * i]->owner,
4517 ovly_sections[2 * i],
4518 ovly_sections[2 * i + 1] ? " + rodata" : "");
4519 bfd_set_error (bfd_error_bad_value);
4520 goto err_exit;
4521 }
4522
4523 while (dummy_caller.call_list != NULL)
4524 {
4525 struct call_info *call = dummy_caller.call_list;
4526 dummy_caller.call_list = call->next;
4527 free (call);
4528 }
4529
4530 ++ovlynum;
4531 while (base < i)
4532 ovly_map[base++] = ovlynum;
4533 }
4534
4535 script = htab->params->spu_elf_open_overlay_script ();
4536
4537 if (htab->params->ovly_flavour == ovly_soft_icache)
4538 {
4539 if (fprintf (script, "SECTIONS\n{\n") <= 0)
4540 goto file_err;
4541
4542 if (fprintf (script,
4543 " . = ALIGN (%u);\n"
4544 " .ovl.init : { *(.ovl.init) }\n"
4545 " . = ABSOLUTE (ADDR (.ovl.init));\n",
4546 htab->params->line_size) <= 0)
4547 goto file_err;
4548
4549 base = 0;
4550 ovlynum = 1;
4551 while (base < count)
4552 {
4553 unsigned int indx = ovlynum - 1;
4554 unsigned int vma, lma;
4555
4556 vma = (indx & (htab->params->num_lines - 1)) << htab->line_size_log2;
4557 lma = vma + (((indx >> htab->num_lines_log2) + 1) << 18);
4558
4559 if (fprintf (script, " .ovly%u ABSOLUTE (ADDR (.ovl.init)) + %u "
4560 ": AT (LOADADDR (.ovl.init) + %u) {\n",
4561 ovlynum, vma, lma) <= 0)
4562 goto file_err;
4563
4564 base = print_one_overlay_section (script, base, count, ovlynum,
4565 ovly_map, ovly_sections, info);
4566 if (base == (unsigned) -1)
4567 goto file_err;
4568
4569 if (fprintf (script, " }\n") <= 0)
4570 goto file_err;
4571
4572 ovlynum++;
4573 }
4574
4575 if (fprintf (script, " . = ABSOLUTE (ADDR (.ovl.init)) + %u;\n",
4576 1 << (htab->num_lines_log2 + htab->line_size_log2)) <= 0)
4577 goto file_err;
4578
4579 if (fprintf (script, "}\nINSERT AFTER .toe;\n") <= 0)
4580 goto file_err;
4581 }
4582 else
4583 {
4584 if (fprintf (script, "SECTIONS\n{\n") <= 0)
4585 goto file_err;
4586
4587 if (fprintf (script,
4588 " . = ALIGN (16);\n"
4589 " .ovl.init : { *(.ovl.init) }\n"
4590 " . = ABSOLUTE (ADDR (.ovl.init));\n") <= 0)
4591 goto file_err;
4592
4593 for (region = 1; region <= htab->params->num_lines; region++)
4594 {
4595 ovlynum = region;
4596 base = 0;
4597 while (base < count && ovly_map[base] < ovlynum)
4598 base++;
4599
4600 if (base == count)
4601 break;
4602
4603 if (region == 1)
4604 {
4605 /* We need to set lma since we are overlaying .ovl.init. */
4606 if (fprintf (script,
4607 " OVERLAY : AT (ALIGN (LOADADDR (.ovl.init) + SIZEOF (.ovl.init), 16))\n {\n") <= 0)
4608 goto file_err;
4609 }
4610 else
4611 {
4612 if (fprintf (script, " OVERLAY :\n {\n") <= 0)
4613 goto file_err;
4614 }
4615
4616 while (base < count)
4617 {
4618 if (fprintf (script, " .ovly%u {\n", ovlynum) <= 0)
4619 goto file_err;
4620
4621 base = print_one_overlay_section (script, base, count, ovlynum,
4622 ovly_map, ovly_sections, info);
4623 if (base == (unsigned) -1)
4624 goto file_err;
4625
4626 if (fprintf (script, " }\n") <= 0)
4627 goto file_err;
4628
4629 ovlynum += htab->params->num_lines;
4630 while (base < count && ovly_map[base] < ovlynum)
4631 base++;
4632 }
4633
4634 if (fprintf (script, " }\n") <= 0)
4635 goto file_err;
4636 }
4637
4638 if (fprintf (script, "}\nINSERT BEFORE .text;\n") <= 0)
4639 goto file_err;
4640 }
4641
4642 free (ovly_map);
4643 free (ovly_sections);
4644
4645 if (fclose (script) != 0)
4646 goto file_err;
4647
4648 if (htab->params->auto_overlay & AUTO_RELINK)
4649 (*htab->params->spu_elf_relink) ();
4650
4651 xexit (0);
4652
4653 file_err:
4654 bfd_set_error (bfd_error_system_call);
4655 err_exit:
4656 info->callbacks->einfo ("%F%P: auto overlay error: %E\n");
4657 xexit (1);
4658 }
4659
4660 /* Provide an estimate of total stack required. */
4661
4662 static bfd_boolean
4663 spu_elf_stack_analysis (struct bfd_link_info *info)
4664 {
4665 struct spu_link_hash_table *htab;
4666 struct _sum_stack_param sum_stack_param;
4667
4668 if (!discover_functions (info))
4669 return FALSE;
4670
4671 if (!build_call_tree (info))
4672 return FALSE;
4673
4674 htab = spu_hash_table (info);
4675 if (htab->params->stack_analysis)
4676 {
4677 info->callbacks->info (_("Stack size for call graph root nodes.\n"));
4678 info->callbacks->minfo (_("\nStack size for functions. "
4679 "Annotations: '*' max stack, 't' tail call\n"));
4680 }
4681
4682 sum_stack_param.emit_stack_syms = htab->params->emit_stack_syms;
4683 sum_stack_param.overall_stack = 0;
4684 if (!for_each_node (sum_stack, info, &sum_stack_param, TRUE))
4685 return FALSE;
4686
4687 if (htab->params->stack_analysis)
4688 info->callbacks->info (_("Maximum stack required is 0x%v\n"),
4689 (bfd_vma) sum_stack_param.overall_stack);
4690 return TRUE;
4691 }
4692
4693 /* Perform a final link. */
4694
4695 static bfd_boolean
4696 spu_elf_final_link (bfd *output_bfd, struct bfd_link_info *info)
4697 {
4698 struct spu_link_hash_table *htab = spu_hash_table (info);
4699
4700 if (htab->params->auto_overlay)
4701 spu_elf_auto_overlay (info);
4702
4703 if ((htab->params->stack_analysis
4704 || (htab->params->ovly_flavour == ovly_soft_icache
4705 && htab->params->lrlive_analysis))
4706 && !spu_elf_stack_analysis (info))
4707 info->callbacks->einfo ("%X%P: stack/lrlive analysis error: %E\n");
4708
4709 if (!spu_elf_build_stubs (info))
4710 info->callbacks->einfo ("%F%P: can not build overlay stubs: %E\n");
4711
4712 return bfd_elf_final_link (output_bfd, info);
4713 }
4714
4715 /* Called when not normally emitting relocs, ie. !info->relocatable
4716 and !info->emitrelocations. Returns a count of special relocs
4717 that need to be emitted. */
4718
4719 static unsigned int
4720 spu_elf_count_relocs (struct bfd_link_info *info, asection *sec)
4721 {
4722 Elf_Internal_Rela *relocs;
4723 unsigned int count = 0;
4724
4725 relocs = _bfd_elf_link_read_relocs (sec->owner, sec, NULL, NULL,
4726 info->keep_memory);
4727 if (relocs != NULL)
4728 {
4729 Elf_Internal_Rela *rel;
4730 Elf_Internal_Rela *relend = relocs + sec->reloc_count;
4731
4732 for (rel = relocs; rel < relend; rel++)
4733 {
4734 int r_type = ELF32_R_TYPE (rel->r_info);
4735 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
4736 ++count;
4737 }
4738
4739 if (elf_section_data (sec)->relocs != relocs)
4740 free (relocs);
4741 }
4742
4743 return count;
4744 }
4745
4746 /* Functions for adding fixup records to .fixup */
4747
4748 #define FIXUP_RECORD_SIZE 4
4749
4750 #define FIXUP_PUT(output_bfd,htab,index,addr) \
4751 bfd_put_32 (output_bfd, addr, \
4752 htab->sfixup->contents + FIXUP_RECORD_SIZE * (index))
4753 #define FIXUP_GET(output_bfd,htab,index) \
4754 bfd_get_32 (output_bfd, \
4755 htab->sfixup->contents + FIXUP_RECORD_SIZE * (index))
4756
4757 /* Store OFFSET in .fixup. This assumes it will be called with an
4758 increasing OFFSET. When this OFFSET fits with the last base offset,
4759 it just sets a bit, otherwise it adds a new fixup record. */
4760 static void
4761 spu_elf_emit_fixup (bfd * output_bfd, struct bfd_link_info *info,
4762 bfd_vma offset)
4763 {
4764 struct spu_link_hash_table *htab = spu_hash_table (info);
4765 asection *sfixup = htab->sfixup;
4766 bfd_vma qaddr = offset & ~(bfd_vma) 15;
4767 bfd_vma bit = ((bfd_vma) 8) >> ((offset & 15) >> 2);
4768 if (sfixup->reloc_count == 0)
4769 {
4770 FIXUP_PUT (output_bfd, htab, 0, qaddr | bit);
4771 sfixup->reloc_count++;
4772 }
4773 else
4774 {
4775 bfd_vma base = FIXUP_GET (output_bfd, htab, sfixup->reloc_count - 1);
4776 if (qaddr != (base & ~(bfd_vma) 15))
4777 {
4778 if ((sfixup->reloc_count + 1) * FIXUP_RECORD_SIZE > sfixup->size)
4779 (*_bfd_error_handler) (_("fatal error while creating .fixup"));
4780 FIXUP_PUT (output_bfd, htab, sfixup->reloc_count, qaddr | bit);
4781 sfixup->reloc_count++;
4782 }
4783 else
4784 FIXUP_PUT (output_bfd, htab, sfixup->reloc_count - 1, base | bit);
4785 }
4786 }
4787
4788 /* Apply RELOCS to CONTENTS of INPUT_SECTION from INPUT_BFD. */
4789
4790 static int
4791 spu_elf_relocate_section (bfd *output_bfd,
4792 struct bfd_link_info *info,
4793 bfd *input_bfd,
4794 asection *input_section,
4795 bfd_byte *contents,
4796 Elf_Internal_Rela *relocs,
4797 Elf_Internal_Sym *local_syms,
4798 asection **local_sections)
4799 {
4800 Elf_Internal_Shdr *symtab_hdr;
4801 struct elf_link_hash_entry **sym_hashes;
4802 Elf_Internal_Rela *rel, *relend;
4803 struct spu_link_hash_table *htab;
4804 asection *ea;
4805 int ret = TRUE;
4806 bfd_boolean emit_these_relocs = FALSE;
4807 bfd_boolean is_ea_sym;
4808 bfd_boolean stubs;
4809 unsigned int iovl = 0;
4810
4811 htab = spu_hash_table (info);
4812 stubs = (htab->stub_sec != NULL
4813 && maybe_needs_stubs (input_section));
4814 iovl = overlay_index (input_section);
4815 ea = bfd_get_section_by_name (output_bfd, "._ea");
4816 symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
4817 sym_hashes = (struct elf_link_hash_entry **) (elf_sym_hashes (input_bfd));
4818
4819 rel = relocs;
4820 relend = relocs + input_section->reloc_count;
4821 for (; rel < relend; rel++)
4822 {
4823 int r_type;
4824 reloc_howto_type *howto;
4825 unsigned int r_symndx;
4826 Elf_Internal_Sym *sym;
4827 asection *sec;
4828 struct elf_link_hash_entry *h;
4829 const char *sym_name;
4830 bfd_vma relocation;
4831 bfd_vma addend;
4832 bfd_reloc_status_type r;
4833 bfd_boolean unresolved_reloc;
4834 enum _stub_type stub_type;
4835
4836 r_symndx = ELF32_R_SYM (rel->r_info);
4837 r_type = ELF32_R_TYPE (rel->r_info);
4838 howto = elf_howto_table + r_type;
4839 unresolved_reloc = FALSE;
4840 h = NULL;
4841 sym = NULL;
4842 sec = NULL;
4843 if (r_symndx < symtab_hdr->sh_info)
4844 {
4845 sym = local_syms + r_symndx;
4846 sec = local_sections[r_symndx];
4847 sym_name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym, sec);
4848 relocation = _bfd_elf_rela_local_sym (output_bfd, sym, &sec, rel);
4849 }
4850 else
4851 {
4852 if (sym_hashes == NULL)
4853 return FALSE;
4854
4855 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
4856
4857 while (h->root.type == bfd_link_hash_indirect
4858 || h->root.type == bfd_link_hash_warning)
4859 h = (struct elf_link_hash_entry *) h->root.u.i.link;
4860
4861 relocation = 0;
4862 if (h->root.type == bfd_link_hash_defined
4863 || h->root.type == bfd_link_hash_defweak)
4864 {
4865 sec = h->root.u.def.section;
4866 if (sec == NULL
4867 || sec->output_section == NULL)
4868 /* Set a flag that will be cleared later if we find a
4869 relocation value for this symbol. output_section
4870 is typically NULL for symbols satisfied by a shared
4871 library. */
4872 unresolved_reloc = TRUE;
4873 else
4874 relocation = (h->root.u.def.value
4875 + sec->output_section->vma
4876 + sec->output_offset);
4877 }
4878 else if (h->root.type == bfd_link_hash_undefweak)
4879 ;
4880 else if (info->unresolved_syms_in_objects == RM_IGNORE
4881 && ELF_ST_VISIBILITY (h->other) == STV_DEFAULT)
4882 ;
4883 else if (!info->relocatable
4884 && !(r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64))
4885 {
4886 bfd_boolean err;
4887 err = (info->unresolved_syms_in_objects == RM_GENERATE_ERROR
4888 || ELF_ST_VISIBILITY (h->other) != STV_DEFAULT);
4889 if (!info->callbacks->undefined_symbol (info,
4890 h->root.root.string,
4891 input_bfd,
4892 input_section,
4893 rel->r_offset, err))
4894 return FALSE;
4895 }
4896 sym_name = h->root.root.string;
4897 }
4898
4899 if (sec != NULL && elf_discarded_section (sec))
4900 RELOC_AGAINST_DISCARDED_SECTION (info, input_bfd, input_section,
4901 rel, relend, howto, contents);
4902
4903 if (info->relocatable)
4904 continue;
4905
4906 /* Change "a rt,ra,rb" to "ai rt,ra,0". */
4907 if (r_type == R_SPU_ADD_PIC
4908 && h != NULL
4909 && !(h->def_regular || ELF_COMMON_DEF_P (h)))
4910 {
4911 bfd_byte *loc = contents + rel->r_offset;
4912 loc[0] = 0x1c;
4913 loc[1] = 0x00;
4914 loc[2] &= 0x3f;
4915 }
4916
4917 is_ea_sym = (ea != NULL
4918 && sec != NULL
4919 && sec->output_section == ea);
4920
4921 /* If this symbol is in an overlay area, we may need to relocate
4922 to the overlay stub. */
4923 addend = rel->r_addend;
4924 if (stubs
4925 && !is_ea_sym
4926 && (stub_type = needs_ovl_stub (h, sym, sec, input_section, rel,
4927 contents, info)) != no_stub)
4928 {
4929 unsigned int ovl = 0;
4930 struct got_entry *g, **head;
4931
4932 if (stub_type != nonovl_stub)
4933 ovl = iovl;
4934
4935 if (h != NULL)
4936 head = &h->got.glist;
4937 else
4938 head = elf_local_got_ents (input_bfd) + r_symndx;
4939
4940 for (g = *head; g != NULL; g = g->next)
4941 if (htab->params->ovly_flavour == ovly_soft_icache
4942 ? (g->ovl == ovl
4943 && g->br_addr == (rel->r_offset
4944 + input_section->output_offset
4945 + input_section->output_section->vma))
4946 : g->addend == addend && (g->ovl == ovl || g->ovl == 0))
4947 break;
4948 if (g == NULL)
4949 abort ();
4950
4951 relocation = g->stub_addr;
4952 addend = 0;
4953 }
4954 else
4955 {
4956 /* For soft icache, encode the overlay index into addresses. */
4957 if (htab->params->ovly_flavour == ovly_soft_icache
4958 && (r_type == R_SPU_ADDR16_HI
4959 || r_type == R_SPU_ADDR32 || r_type == R_SPU_REL32)
4960 && !is_ea_sym)
4961 {
4962 unsigned int ovl = overlay_index (sec);
4963 if (ovl != 0)
4964 {
4965 unsigned int set_id = ((ovl - 1) >> htab->num_lines_log2) + 1;
4966 relocation += set_id << 18;
4967 }
4968 }
4969 }
4970
4971 if (htab->params->emit_fixups && !info->relocatable
4972 && (input_section->flags & SEC_ALLOC) != 0
4973 && r_type == R_SPU_ADDR32)
4974 {
4975 bfd_vma offset;
4976 offset = rel->r_offset + input_section->output_section->vma
4977 + input_section->output_offset;
4978 spu_elf_emit_fixup (output_bfd, info, offset);
4979 }
4980
4981 if (unresolved_reloc)
4982 ;
4983 else if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
4984 {
4985 if (is_ea_sym)
4986 {
4987 /* ._ea is a special section that isn't allocated in SPU
4988 memory, but rather occupies space in PPU memory as
4989 part of an embedded ELF image. If this reloc is
4990 against a symbol defined in ._ea, then transform the
4991 reloc into an equivalent one without a symbol
4992 relative to the start of the ELF image. */
4993 rel->r_addend += (relocation
4994 - ea->vma
4995 + elf_section_data (ea)->this_hdr.sh_offset);
4996 rel->r_info = ELF32_R_INFO (0, r_type);
4997 }
4998 emit_these_relocs = TRUE;
4999 continue;
5000 }
5001 else if (is_ea_sym)
5002 unresolved_reloc = TRUE;
5003
5004 if (unresolved_reloc
5005 && _bfd_elf_section_offset (output_bfd, info, input_section,
5006 rel->r_offset) != (bfd_vma) -1)
5007 {
5008 (*_bfd_error_handler)
5009 (_("%B(%s+0x%lx): unresolvable %s relocation against symbol `%s'"),
5010 input_bfd,
5011 bfd_get_section_name (input_bfd, input_section),
5012 (long) rel->r_offset,
5013 howto->name,
5014 sym_name);
5015 ret = FALSE;
5016 }
5017
5018 r = _bfd_final_link_relocate (howto,
5019 input_bfd,
5020 input_section,
5021 contents,
5022 rel->r_offset, relocation, addend);
5023
5024 if (r != bfd_reloc_ok)
5025 {
5026 const char *msg = (const char *) 0;
5027
5028 switch (r)
5029 {
5030 case bfd_reloc_overflow:
5031 if (!((*info->callbacks->reloc_overflow)
5032 (info, (h ? &h->root : NULL), sym_name, howto->name,
5033 (bfd_vma) 0, input_bfd, input_section, rel->r_offset)))
5034 return FALSE;
5035 break;
5036
5037 case bfd_reloc_undefined:
5038 if (!((*info->callbacks->undefined_symbol)
5039 (info, sym_name, input_bfd, input_section,
5040 rel->r_offset, TRUE)))
5041 return FALSE;
5042 break;
5043
5044 case bfd_reloc_outofrange:
5045 msg = _("internal error: out of range error");
5046 goto common_error;
5047
5048 case bfd_reloc_notsupported:
5049 msg = _("internal error: unsupported relocation error");
5050 goto common_error;
5051
5052 case bfd_reloc_dangerous:
5053 msg = _("internal error: dangerous error");
5054 goto common_error;
5055
5056 default:
5057 msg = _("internal error: unknown error");
5058 /* fall through */
5059
5060 common_error:
5061 ret = FALSE;
5062 if (!((*info->callbacks->warning)
5063 (info, msg, sym_name, input_bfd, input_section,
5064 rel->r_offset)))
5065 return FALSE;
5066 break;
5067 }
5068 }
5069 }
5070
5071 if (ret
5072 && emit_these_relocs
5073 && !info->emitrelocations)
5074 {
5075 Elf_Internal_Rela *wrel;
5076 Elf_Internal_Shdr *rel_hdr;
5077
5078 wrel = rel = relocs;
5079 relend = relocs + input_section->reloc_count;
5080 for (; rel < relend; rel++)
5081 {
5082 int r_type;
5083
5084 r_type = ELF32_R_TYPE (rel->r_info);
5085 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
5086 *wrel++ = *rel;
5087 }
5088 input_section->reloc_count = wrel - relocs;
5089 /* Backflips for _bfd_elf_link_output_relocs. */
5090 rel_hdr = _bfd_elf_single_rel_hdr (input_section);
5091 rel_hdr->sh_size = input_section->reloc_count * rel_hdr->sh_entsize;
5092 ret = 2;
5093 }
5094
5095 return ret;
5096 }
5097
5098 static bfd_boolean
5099 spu_elf_finish_dynamic_sections (bfd *output_bfd ATTRIBUTE_UNUSED,
5100 struct bfd_link_info *info ATTRIBUTE_UNUSED)
5101 {
5102 return TRUE;
5103 }
5104
5105 /* Adjust _SPUEAR_ syms to point at their overlay stubs. */
5106
5107 static int
5108 spu_elf_output_symbol_hook (struct bfd_link_info *info,
5109 const char *sym_name ATTRIBUTE_UNUSED,
5110 Elf_Internal_Sym *sym,
5111 asection *sym_sec ATTRIBUTE_UNUSED,
5112 struct elf_link_hash_entry *h)
5113 {
5114 struct spu_link_hash_table *htab = spu_hash_table (info);
5115
5116 if (!info->relocatable
5117 && htab->stub_sec != NULL
5118 && h != NULL
5119 && (h->root.type == bfd_link_hash_defined
5120 || h->root.type == bfd_link_hash_defweak)
5121 && h->def_regular
5122 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0)
5123 {
5124 struct got_entry *g;
5125
5126 for (g = h->got.glist; g != NULL; g = g->next)
5127 if (htab->params->ovly_flavour == ovly_soft_icache
5128 ? g->br_addr == g->stub_addr
5129 : g->addend == 0 && g->ovl == 0)
5130 {
5131 sym->st_shndx = (_bfd_elf_section_from_bfd_section
5132 (htab->stub_sec[0]->output_section->owner,
5133 htab->stub_sec[0]->output_section));
5134 sym->st_value = g->stub_addr;
5135 break;
5136 }
5137 }
5138
5139 return 1;
5140 }
5141
5142 static int spu_plugin = 0;
5143
5144 void
5145 spu_elf_plugin (int val)
5146 {
5147 spu_plugin = val;
5148 }
5149
5150 /* Set ELF header e_type for plugins. */
5151
5152 static void
5153 spu_elf_post_process_headers (bfd *abfd,
5154 struct bfd_link_info *info ATTRIBUTE_UNUSED)
5155 {
5156 if (spu_plugin)
5157 {
5158 Elf_Internal_Ehdr *i_ehdrp = elf_elfheader (abfd);
5159
5160 i_ehdrp->e_type = ET_DYN;
5161 }
5162 }
5163
5164 /* We may add an extra PT_LOAD segment for .toe. We also need extra
5165 segments for overlays. */
5166
5167 static int
5168 spu_elf_additional_program_headers (bfd *abfd, struct bfd_link_info *info)
5169 {
5170 int extra = 0;
5171 asection *sec;
5172
5173 if (info != NULL)
5174 {
5175 struct spu_link_hash_table *htab = spu_hash_table (info);
5176 extra = htab->num_overlays;
5177 }
5178
5179 if (extra)
5180 ++extra;
5181
5182 sec = bfd_get_section_by_name (abfd, ".toe");
5183 if (sec != NULL && (sec->flags & SEC_LOAD) != 0)
5184 ++extra;
5185
5186 return extra;
5187 }
5188
5189 /* Remove .toe section from other PT_LOAD segments and put it in
5190 a segment of its own. Put overlays in separate segments too. */
5191
5192 static bfd_boolean
5193 spu_elf_modify_segment_map (bfd *abfd, struct bfd_link_info *info)
5194 {
5195 asection *toe, *s;
5196 struct elf_segment_map *m, *m_overlay;
5197 struct elf_segment_map **p, **p_overlay;
5198 unsigned int i;
5199
5200 if (info == NULL)
5201 return TRUE;
5202
5203 toe = bfd_get_section_by_name (abfd, ".toe");
5204 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
5205 if (m->p_type == PT_LOAD && m->count > 1)
5206 for (i = 0; i < m->count; i++)
5207 if ((s = m->sections[i]) == toe
5208 || spu_elf_section_data (s)->u.o.ovl_index != 0)
5209 {
5210 struct elf_segment_map *m2;
5211 bfd_vma amt;
5212
5213 if (i + 1 < m->count)
5214 {
5215 amt = sizeof (struct elf_segment_map);
5216 amt += (m->count - (i + 2)) * sizeof (m->sections[0]);
5217 m2 = bfd_zalloc (abfd, amt);
5218 if (m2 == NULL)
5219 return FALSE;
5220 m2->count = m->count - (i + 1);
5221 memcpy (m2->sections, m->sections + i + 1,
5222 m2->count * sizeof (m->sections[0]));
5223 m2->p_type = PT_LOAD;
5224 m2->next = m->next;
5225 m->next = m2;
5226 }
5227 m->count = 1;
5228 if (i != 0)
5229 {
5230 m->count = i;
5231 amt = sizeof (struct elf_segment_map);
5232 m2 = bfd_zalloc (abfd, amt);
5233 if (m2 == NULL)
5234 return FALSE;
5235 m2->p_type = PT_LOAD;
5236 m2->count = 1;
5237 m2->sections[0] = s;
5238 m2->next = m->next;
5239 m->next = m2;
5240 }
5241 break;
5242 }
5243
5244
5245 /* Some SPU ELF loaders ignore the PF_OVERLAY flag and just load all
5246 PT_LOAD segments. This can cause the .ovl.init section to be
5247 overwritten with the contents of some overlay segment. To work
5248 around this issue, we ensure that all PF_OVERLAY segments are
5249 sorted first amongst the program headers; this ensures that even
5250 with a broken loader, the .ovl.init section (which is not marked
5251 as PF_OVERLAY) will be placed into SPU local store on startup. */
5252
5253 /* Move all overlay segments onto a separate list. */
5254 p = &elf_tdata (abfd)->segment_map;
5255 p_overlay = &m_overlay;
5256 while (*p != NULL)
5257 {
5258 if ((*p)->p_type == PT_LOAD && (*p)->count == 1
5259 && spu_elf_section_data ((*p)->sections[0])->u.o.ovl_index != 0)
5260 {
5261 m = *p;
5262 *p = m->next;
5263 *p_overlay = m;
5264 p_overlay = &m->next;
5265 continue;
5266 }
5267
5268 p = &((*p)->next);
5269 }
5270
5271 /* Re-insert overlay segments at the head of the segment map. */
5272 *p_overlay = elf_tdata (abfd)->segment_map;
5273 elf_tdata (abfd)->segment_map = m_overlay;
5274
5275 return TRUE;
5276 }
5277
5278 /* Tweak the section type of .note.spu_name. */
5279
5280 static bfd_boolean
5281 spu_elf_fake_sections (bfd *obfd ATTRIBUTE_UNUSED,
5282 Elf_Internal_Shdr *hdr,
5283 asection *sec)
5284 {
5285 if (strcmp (sec->name, SPU_PTNOTE_SPUNAME) == 0)
5286 hdr->sh_type = SHT_NOTE;
5287 return TRUE;
5288 }
5289
5290 /* Tweak phdrs before writing them out. */
5291
5292 static int
5293 spu_elf_modify_program_headers (bfd *abfd, struct bfd_link_info *info)
5294 {
5295 const struct elf_backend_data *bed;
5296 struct elf_obj_tdata *tdata;
5297 Elf_Internal_Phdr *phdr, *last;
5298 struct spu_link_hash_table *htab;
5299 unsigned int count;
5300 unsigned int i;
5301
5302 if (info == NULL)
5303 return TRUE;
5304
5305 bed = get_elf_backend_data (abfd);
5306 tdata = elf_tdata (abfd);
5307 phdr = tdata->phdr;
5308 count = tdata->program_header_size / bed->s->sizeof_phdr;
5309 htab = spu_hash_table (info);
5310 if (htab->num_overlays != 0)
5311 {
5312 struct elf_segment_map *m;
5313 unsigned int o;
5314
5315 for (i = 0, m = elf_tdata (abfd)->segment_map; m; ++i, m = m->next)
5316 if (m->count != 0
5317 && (o = spu_elf_section_data (m->sections[0])->u.o.ovl_index) != 0)
5318 {
5319 /* Mark this as an overlay header. */
5320 phdr[i].p_flags |= PF_OVERLAY;
5321
5322 if (htab->ovtab != NULL && htab->ovtab->size != 0
5323 && htab->params->ovly_flavour != ovly_soft_icache)
5324 {
5325 bfd_byte *p = htab->ovtab->contents;
5326 unsigned int off = o * 16 + 8;
5327
5328 /* Write file_off into _ovly_table. */
5329 bfd_put_32 (htab->ovtab->owner, phdr[i].p_offset, p + off);
5330 }
5331 }
5332 /* Soft-icache has its file offset put in .ovl.init. */
5333 if (htab->init != NULL && htab->init->size != 0)
5334 {
5335 bfd_vma val = elf_section_data (htab->ovl_sec[0])->this_hdr.sh_offset;
5336
5337 bfd_put_32 (htab->init->owner, val, htab->init->contents + 4);
5338 }
5339 }
5340
5341 /* Round up p_filesz and p_memsz of PT_LOAD segments to multiples
5342 of 16. This should always be possible when using the standard
5343 linker scripts, but don't create overlapping segments if
5344 someone is playing games with linker scripts. */
5345 last = NULL;
5346 for (i = count; i-- != 0; )
5347 if (phdr[i].p_type == PT_LOAD)
5348 {
5349 unsigned adjust;
5350
5351 adjust = -phdr[i].p_filesz & 15;
5352 if (adjust != 0
5353 && last != NULL
5354 && phdr[i].p_offset + phdr[i].p_filesz > last->p_offset - adjust)
5355 break;
5356
5357 adjust = -phdr[i].p_memsz & 15;
5358 if (adjust != 0
5359 && last != NULL
5360 && phdr[i].p_filesz != 0
5361 && phdr[i].p_vaddr + phdr[i].p_memsz > last->p_vaddr - adjust
5362 && phdr[i].p_vaddr + phdr[i].p_memsz <= last->p_vaddr)
5363 break;
5364
5365 if (phdr[i].p_filesz != 0)
5366 last = &phdr[i];
5367 }
5368
5369 if (i == (unsigned int) -1)
5370 for (i = count; i-- != 0; )
5371 if (phdr[i].p_type == PT_LOAD)
5372 {
5373 unsigned adjust;
5374
5375 adjust = -phdr[i].p_filesz & 15;
5376 phdr[i].p_filesz += adjust;
5377
5378 adjust = -phdr[i].p_memsz & 15;
5379 phdr[i].p_memsz += adjust;
5380 }
5381
5382 return TRUE;
5383 }
5384
5385 bfd_boolean
5386 spu_elf_size_sections (bfd * output_bfd, struct bfd_link_info *info)
5387 {
5388 struct spu_link_hash_table *htab = spu_hash_table (info);
5389 if (htab->params->emit_fixups)
5390 {
5391 asection *sfixup = htab->sfixup;
5392 int fixup_count = 0;
5393 bfd *ibfd;
5394 size_t size;
5395
5396 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
5397 {
5398 asection *isec;
5399
5400 if (bfd_get_flavour (ibfd) != bfd_target_elf_flavour)
5401 continue;
5402
5403 /* Walk over each section attached to the input bfd. */
5404 for (isec = ibfd->sections; isec != NULL; isec = isec->next)
5405 {
5406 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
5407 bfd_vma base_end;
5408
5409 /* If there aren't any relocs, then there's nothing more
5410 to do. */
5411 if ((isec->flags & SEC_ALLOC) == 0
5412 || (isec->flags & SEC_RELOC) == 0
5413 || isec->reloc_count == 0)
5414 continue;
5415
5416 /* Get the relocs. */
5417 internal_relocs =
5418 _bfd_elf_link_read_relocs (ibfd, isec, NULL, NULL,
5419 info->keep_memory);
5420 if (internal_relocs == NULL)
5421 return FALSE;
5422
5423 /* 1 quadword can contain up to 4 R_SPU_ADDR32
5424 relocations. They are stored in a single word by
5425 saving the upper 28 bits of the address and setting the
5426 lower 4 bits to a bit mask of the words that have the
5427 relocation. BASE_END keeps track of the next quadword. */
5428 irela = internal_relocs;
5429 irelaend = irela + isec->reloc_count;
5430 base_end = 0;
5431 for (; irela < irelaend; irela++)
5432 if (ELF32_R_TYPE (irela->r_info) == R_SPU_ADDR32
5433 && irela->r_offset >= base_end)
5434 {
5435 base_end = (irela->r_offset & ~(bfd_vma) 15) + 16;
5436 fixup_count++;
5437 }
5438 }
5439 }
5440
5441 /* We always have a NULL fixup as a sentinel */
5442 size = (fixup_count + 1) * FIXUP_RECORD_SIZE;
5443 if (!bfd_set_section_size (output_bfd, sfixup, size))
5444 return FALSE;
5445 sfixup->contents = (bfd_byte *) bfd_zalloc (info->input_bfds, size);
5446 if (sfixup->contents == NULL)
5447 return FALSE;
5448 }
5449 return TRUE;
5450 }
5451
5452 #define TARGET_BIG_SYM bfd_elf32_spu_vec
5453 #define TARGET_BIG_NAME "elf32-spu"
5454 #define ELF_ARCH bfd_arch_spu
5455 #define ELF_TARGET_ID SPU_ELF_DATA
5456 #define ELF_MACHINE_CODE EM_SPU
5457 /* This matches the alignment need for DMA. */
5458 #define ELF_MAXPAGESIZE 0x80
5459 #define elf_backend_rela_normal 1
5460 #define elf_backend_can_gc_sections 1
5461
5462 #define bfd_elf32_bfd_reloc_type_lookup spu_elf_reloc_type_lookup
5463 #define bfd_elf32_bfd_reloc_name_lookup spu_elf_reloc_name_lookup
5464 #define elf_info_to_howto spu_elf_info_to_howto
5465 #define elf_backend_count_relocs spu_elf_count_relocs
5466 #define elf_backend_relocate_section spu_elf_relocate_section
5467 #define elf_backend_finish_dynamic_sections spu_elf_finish_dynamic_sections
5468 #define elf_backend_symbol_processing spu_elf_backend_symbol_processing
5469 #define elf_backend_link_output_symbol_hook spu_elf_output_symbol_hook
5470 #define elf_backend_object_p spu_elf_object_p
5471 #define bfd_elf32_new_section_hook spu_elf_new_section_hook
5472 #define bfd_elf32_bfd_link_hash_table_create spu_elf_link_hash_table_create
5473
5474 #define elf_backend_additional_program_headers spu_elf_additional_program_headers
5475 #define elf_backend_modify_segment_map spu_elf_modify_segment_map
5476 #define elf_backend_modify_program_headers spu_elf_modify_program_headers
5477 #define elf_backend_post_process_headers spu_elf_post_process_headers
5478 #define elf_backend_fake_sections spu_elf_fake_sections
5479 #define elf_backend_special_sections spu_elf_special_sections
5480 #define bfd_elf32_bfd_final_link spu_elf_final_link
5481
5482 #include "elf32-target.h"
This page took 0.219956 seconds and 4 git commands to generate.