2008-01-30 Ralf Wildenhues <Ralf.Wildenhues@gmx.de>
[binutils.git] / bfd / elf32-spu.c
blobd9ba935c28956d4a1356cf52be9d6642c7eb1726
1 /* SPU specific support for 32-bit ELF
3 Copyright 2006, 2007 Free Software Foundation, Inc.
5 This file is part of BFD, the Binary File Descriptor library.
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3 of the License, or
10 (at your option) any later version.
12 This program is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License along
18 with this program; if not, write to the Free Software Foundation, Inc.,
19 51 Franklin Street - Fifth Floor, Boston, MA 02110-1301, USA. */
21 #include "sysdep.h"
22 #include "bfd.h"
23 #include "bfdlink.h"
24 #include "libbfd.h"
25 #include "elf-bfd.h"
26 #include "elf/spu.h"
27 #include "elf32-spu.h"
29 /* We use RELA style relocs. Don't define USE_REL. */
31 static bfd_reloc_status_type spu_elf_rel9 (bfd *, arelent *, asymbol *,
32 void *, asection *,
33 bfd *, char **);
35 /* Values of type 'enum elf_spu_reloc_type' are used to index this
36 array, so it must be declared in the order of that type. */
38 static reloc_howto_type elf_howto_table[] = {
39 HOWTO (R_SPU_NONE, 0, 0, 0, FALSE, 0, complain_overflow_dont,
40 bfd_elf_generic_reloc, "SPU_NONE",
41 FALSE, 0, 0x00000000, FALSE),
42 HOWTO (R_SPU_ADDR10, 4, 2, 10, FALSE, 14, complain_overflow_bitfield,
43 bfd_elf_generic_reloc, "SPU_ADDR10",
44 FALSE, 0, 0x00ffc000, FALSE),
45 HOWTO (R_SPU_ADDR16, 2, 2, 16, FALSE, 7, complain_overflow_bitfield,
46 bfd_elf_generic_reloc, "SPU_ADDR16",
47 FALSE, 0, 0x007fff80, FALSE),
48 HOWTO (R_SPU_ADDR16_HI, 16, 2, 16, FALSE, 7, complain_overflow_bitfield,
49 bfd_elf_generic_reloc, "SPU_ADDR16_HI",
50 FALSE, 0, 0x007fff80, FALSE),
51 HOWTO (R_SPU_ADDR16_LO, 0, 2, 16, FALSE, 7, complain_overflow_dont,
52 bfd_elf_generic_reloc, "SPU_ADDR16_LO",
53 FALSE, 0, 0x007fff80, FALSE),
54 HOWTO (R_SPU_ADDR18, 0, 2, 18, FALSE, 7, complain_overflow_bitfield,
55 bfd_elf_generic_reloc, "SPU_ADDR18",
56 FALSE, 0, 0x01ffff80, FALSE),
57 HOWTO (R_SPU_ADDR32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
58 bfd_elf_generic_reloc, "SPU_ADDR32",
59 FALSE, 0, 0xffffffff, FALSE),
60 HOWTO (R_SPU_REL16, 2, 2, 16, TRUE, 7, complain_overflow_bitfield,
61 bfd_elf_generic_reloc, "SPU_REL16",
62 FALSE, 0, 0x007fff80, TRUE),
63 HOWTO (R_SPU_ADDR7, 0, 2, 7, FALSE, 14, complain_overflow_dont,
64 bfd_elf_generic_reloc, "SPU_ADDR7",
65 FALSE, 0, 0x001fc000, FALSE),
66 HOWTO (R_SPU_REL9, 2, 2, 9, TRUE, 0, complain_overflow_signed,
67 spu_elf_rel9, "SPU_REL9",
68 FALSE, 0, 0x0180007f, TRUE),
69 HOWTO (R_SPU_REL9I, 2, 2, 9, TRUE, 0, complain_overflow_signed,
70 spu_elf_rel9, "SPU_REL9I",
71 FALSE, 0, 0x0000c07f, TRUE),
72 HOWTO (R_SPU_ADDR10I, 0, 2, 10, FALSE, 14, complain_overflow_signed,
73 bfd_elf_generic_reloc, "SPU_ADDR10I",
74 FALSE, 0, 0x00ffc000, FALSE),
75 HOWTO (R_SPU_ADDR16I, 0, 2, 16, FALSE, 7, complain_overflow_signed,
76 bfd_elf_generic_reloc, "SPU_ADDR16I",
77 FALSE, 0, 0x007fff80, FALSE),
78 HOWTO (R_SPU_REL32, 0, 2, 32, TRUE, 0, complain_overflow_dont,
79 bfd_elf_generic_reloc, "SPU_REL32",
80 FALSE, 0, 0xffffffff, TRUE),
81 HOWTO (R_SPU_ADDR16X, 0, 2, 16, FALSE, 7, complain_overflow_bitfield,
82 bfd_elf_generic_reloc, "SPU_ADDR16X",
83 FALSE, 0, 0x007fff80, FALSE),
84 HOWTO (R_SPU_PPU32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
85 bfd_elf_generic_reloc, "SPU_PPU32",
86 FALSE, 0, 0xffffffff, FALSE),
87 HOWTO (R_SPU_PPU64, 0, 4, 64, FALSE, 0, complain_overflow_dont,
88 bfd_elf_generic_reloc, "SPU_PPU64",
89 FALSE, 0, -1, FALSE),
92 static struct bfd_elf_special_section const spu_elf_special_sections[] = {
93 { ".toe", 4, 0, SHT_NOBITS, SHF_ALLOC },
94 { NULL, 0, 0, 0, 0 }
97 static enum elf_spu_reloc_type
98 spu_elf_bfd_to_reloc_type (bfd_reloc_code_real_type code)
100 switch (code)
102 default:
103 return R_SPU_NONE;
104 case BFD_RELOC_SPU_IMM10W:
105 return R_SPU_ADDR10;
106 case BFD_RELOC_SPU_IMM16W:
107 return R_SPU_ADDR16;
108 case BFD_RELOC_SPU_LO16:
109 return R_SPU_ADDR16_LO;
110 case BFD_RELOC_SPU_HI16:
111 return R_SPU_ADDR16_HI;
112 case BFD_RELOC_SPU_IMM18:
113 return R_SPU_ADDR18;
114 case BFD_RELOC_SPU_PCREL16:
115 return R_SPU_REL16;
116 case BFD_RELOC_SPU_IMM7:
117 return R_SPU_ADDR7;
118 case BFD_RELOC_SPU_IMM8:
119 return R_SPU_NONE;
120 case BFD_RELOC_SPU_PCREL9a:
121 return R_SPU_REL9;
122 case BFD_RELOC_SPU_PCREL9b:
123 return R_SPU_REL9I;
124 case BFD_RELOC_SPU_IMM10:
125 return R_SPU_ADDR10I;
126 case BFD_RELOC_SPU_IMM16:
127 return R_SPU_ADDR16I;
128 case BFD_RELOC_32:
129 return R_SPU_ADDR32;
130 case BFD_RELOC_32_PCREL:
131 return R_SPU_REL32;
132 case BFD_RELOC_SPU_PPU32:
133 return R_SPU_PPU32;
134 case BFD_RELOC_SPU_PPU64:
135 return R_SPU_PPU64;
139 static void
140 spu_elf_info_to_howto (bfd *abfd ATTRIBUTE_UNUSED,
141 arelent *cache_ptr,
142 Elf_Internal_Rela *dst)
144 enum elf_spu_reloc_type r_type;
146 r_type = (enum elf_spu_reloc_type) ELF32_R_TYPE (dst->r_info);
147 BFD_ASSERT (r_type < R_SPU_max);
148 cache_ptr->howto = &elf_howto_table[(int) r_type];
151 static reloc_howto_type *
152 spu_elf_reloc_type_lookup (bfd *abfd ATTRIBUTE_UNUSED,
153 bfd_reloc_code_real_type code)
155 enum elf_spu_reloc_type r_type = spu_elf_bfd_to_reloc_type (code);
157 if (r_type == R_SPU_NONE)
158 return NULL;
160 return elf_howto_table + r_type;
163 static reloc_howto_type *
164 spu_elf_reloc_name_lookup (bfd *abfd ATTRIBUTE_UNUSED,
165 const char *r_name)
167 unsigned int i;
169 for (i = 0; i < sizeof (elf_howto_table) / sizeof (elf_howto_table[0]); i++)
170 if (elf_howto_table[i].name != NULL
171 && strcasecmp (elf_howto_table[i].name, r_name) == 0)
172 return &elf_howto_table[i];
174 return NULL;
177 /* Apply R_SPU_REL9 and R_SPU_REL9I relocs. */
179 static bfd_reloc_status_type
180 spu_elf_rel9 (bfd *abfd, arelent *reloc_entry, asymbol *symbol,
181 void *data, asection *input_section,
182 bfd *output_bfd, char **error_message)
184 bfd_size_type octets;
185 bfd_vma val;
186 long insn;
188 /* If this is a relocatable link (output_bfd test tells us), just
189 call the generic function. Any adjustment will be done at final
190 link time. */
191 if (output_bfd != NULL)
192 return bfd_elf_generic_reloc (abfd, reloc_entry, symbol, data,
193 input_section, output_bfd, error_message);
195 if (reloc_entry->address > bfd_get_section_limit (abfd, input_section))
196 return bfd_reloc_outofrange;
197 octets = reloc_entry->address * bfd_octets_per_byte (abfd);
199 /* Get symbol value. */
200 val = 0;
201 if (!bfd_is_com_section (symbol->section))
202 val = symbol->value;
203 if (symbol->section->output_section)
204 val += symbol->section->output_section->vma;
206 val += reloc_entry->addend;
208 /* Make it pc-relative. */
209 val -= input_section->output_section->vma + input_section->output_offset;
211 val >>= 2;
212 if (val + 256 >= 512)
213 return bfd_reloc_overflow;
215 insn = bfd_get_32 (abfd, (bfd_byte *) data + octets);
217 /* Move two high bits of value to REL9I and REL9 position.
218 The mask will take care of selecting the right field. */
219 val = (val & 0x7f) | ((val & 0x180) << 7) | ((val & 0x180) << 16);
220 insn &= ~reloc_entry->howto->dst_mask;
221 insn |= val & reloc_entry->howto->dst_mask;
222 bfd_put_32 (abfd, insn, (bfd_byte *) data + octets);
223 return bfd_reloc_ok;
226 static bfd_boolean
227 spu_elf_new_section_hook (bfd *abfd, asection *sec)
229 if (!sec->used_by_bfd)
231 struct _spu_elf_section_data *sdata;
233 sdata = bfd_zalloc (abfd, sizeof (*sdata));
234 if (sdata == NULL)
235 return FALSE;
236 sec->used_by_bfd = sdata;
239 return _bfd_elf_new_section_hook (abfd, sec);
242 /* Specially mark defined symbols named _EAR_* with BSF_KEEP so that
243 strip --strip-unneeded will not remove them. */
245 static void
246 spu_elf_backend_symbol_processing (bfd *abfd ATTRIBUTE_UNUSED, asymbol *sym)
248 if (sym->name != NULL
249 && sym->section != bfd_abs_section_ptr
250 && strncmp (sym->name, "_EAR_", 5) == 0)
251 sym->flags |= BSF_KEEP;
254 /* SPU ELF linker hash table. */
256 struct spu_link_hash_table
258 struct elf_link_hash_table elf;
260 /* Shortcuts to overlay sections. */
261 asection *ovtab;
262 asection *toe;
263 asection **ovl_sec;
265 /* Count of stubs in each overlay section. */
266 unsigned int *stub_count;
268 /* The stub section for each overlay section. */
269 asection **stub_sec;
271 struct elf_link_hash_entry *ovly_load;
272 struct elf_link_hash_entry *ovly_return;
273 unsigned long ovly_load_r_symndx;
275 /* Number of overlay buffers. */
276 unsigned int num_buf;
278 /* Total number of overlays. */
279 unsigned int num_overlays;
281 /* Set if we should emit symbols for stubs. */
282 unsigned int emit_stub_syms:1;
284 /* Set if we want stubs on calls out of overlay regions to
285 non-overlay regions. */
286 unsigned int non_overlay_stubs : 1;
288 /* Set on error. */
289 unsigned int stub_err : 1;
291 /* Set if stack size analysis should be done. */
292 unsigned int stack_analysis : 1;
294 /* Set if __stack_* syms will be emitted. */
295 unsigned int emit_stack_syms : 1;
298 /* Hijack the generic got fields for overlay stub accounting. */
300 struct got_entry
302 struct got_entry *next;
303 unsigned int ovl;
304 bfd_vma stub_addr;
307 #define spu_hash_table(p) \
308 ((struct spu_link_hash_table *) ((p)->hash))
310 /* Create a spu ELF linker hash table. */
312 static struct bfd_link_hash_table *
313 spu_elf_link_hash_table_create (bfd *abfd)
315 struct spu_link_hash_table *htab;
317 htab = bfd_malloc (sizeof (*htab));
318 if (htab == NULL)
319 return NULL;
321 if (!_bfd_elf_link_hash_table_init (&htab->elf, abfd,
322 _bfd_elf_link_hash_newfunc,
323 sizeof (struct elf_link_hash_entry)))
325 free (htab);
326 return NULL;
329 memset (&htab->ovtab, 0,
330 sizeof (*htab) - offsetof (struct spu_link_hash_table, ovtab));
332 htab->elf.init_got_refcount.refcount = 0;
333 htab->elf.init_got_refcount.glist = NULL;
334 htab->elf.init_got_offset.offset = 0;
335 htab->elf.init_got_offset.glist = NULL;
336 return &htab->elf.root;
339 /* Find the symbol for the given R_SYMNDX in IBFD and set *HP and *SYMP
340 to (hash, NULL) for global symbols, and (NULL, sym) for locals. Set
341 *SYMSECP to the symbol's section. *LOCSYMSP caches local syms. */
343 static bfd_boolean
344 get_sym_h (struct elf_link_hash_entry **hp,
345 Elf_Internal_Sym **symp,
346 asection **symsecp,
347 Elf_Internal_Sym **locsymsp,
348 unsigned long r_symndx,
349 bfd *ibfd)
351 Elf_Internal_Shdr *symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
353 if (r_symndx >= symtab_hdr->sh_info)
355 struct elf_link_hash_entry **sym_hashes = elf_sym_hashes (ibfd);
356 struct elf_link_hash_entry *h;
358 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
359 while (h->root.type == bfd_link_hash_indirect
360 || h->root.type == bfd_link_hash_warning)
361 h = (struct elf_link_hash_entry *) h->root.u.i.link;
363 if (hp != NULL)
364 *hp = h;
366 if (symp != NULL)
367 *symp = NULL;
369 if (symsecp != NULL)
371 asection *symsec = NULL;
372 if (h->root.type == bfd_link_hash_defined
373 || h->root.type == bfd_link_hash_defweak)
374 symsec = h->root.u.def.section;
375 *symsecp = symsec;
378 else
380 Elf_Internal_Sym *sym;
381 Elf_Internal_Sym *locsyms = *locsymsp;
383 if (locsyms == NULL)
385 locsyms = (Elf_Internal_Sym *) symtab_hdr->contents;
386 if (locsyms == NULL)
388 size_t symcount = symtab_hdr->sh_info;
390 /* If we are reading symbols into the contents, then
391 read the global syms too. This is done to cache
392 syms for later stack analysis. */
393 if ((unsigned char **) locsymsp == &symtab_hdr->contents)
394 symcount = symtab_hdr->sh_size / symtab_hdr->sh_entsize;
395 locsyms = bfd_elf_get_elf_syms (ibfd, symtab_hdr, symcount, 0,
396 NULL, NULL, NULL);
398 if (locsyms == NULL)
399 return FALSE;
400 *locsymsp = locsyms;
402 sym = locsyms + r_symndx;
404 if (hp != NULL)
405 *hp = NULL;
407 if (symp != NULL)
408 *symp = sym;
410 if (symsecp != NULL)
412 asection *symsec = NULL;
413 if ((sym->st_shndx != SHN_UNDEF
414 && sym->st_shndx < SHN_LORESERVE)
415 || sym->st_shndx > SHN_HIRESERVE)
416 symsec = bfd_section_from_elf_index (ibfd, sym->st_shndx);
417 *symsecp = symsec;
421 return TRUE;
424 /* Create the note section if not already present. This is done early so
425 that the linker maps the sections to the right place in the output. */
427 bfd_boolean
428 spu_elf_create_sections (bfd *output_bfd,
429 struct bfd_link_info *info,
430 int stack_analysis,
431 int emit_stack_syms)
433 bfd *ibfd;
434 struct spu_link_hash_table *htab = spu_hash_table (info);
436 /* Stash some options away where we can get at them later. */
437 htab->stack_analysis = stack_analysis;
438 htab->emit_stack_syms = emit_stack_syms;
440 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
441 if (bfd_get_section_by_name (ibfd, SPU_PTNOTE_SPUNAME) != NULL)
442 break;
444 if (ibfd == NULL)
446 /* Make SPU_PTNOTE_SPUNAME section. */
447 asection *s;
448 size_t name_len;
449 size_t size;
450 bfd_byte *data;
451 flagword flags;
453 ibfd = info->input_bfds;
454 flags = SEC_LOAD | SEC_READONLY | SEC_HAS_CONTENTS | SEC_IN_MEMORY;
455 s = bfd_make_section_anyway_with_flags (ibfd, SPU_PTNOTE_SPUNAME, flags);
456 if (s == NULL
457 || !bfd_set_section_alignment (ibfd, s, 4))
458 return FALSE;
460 name_len = strlen (bfd_get_filename (output_bfd)) + 1;
461 size = 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4);
462 size += (name_len + 3) & -4;
464 if (!bfd_set_section_size (ibfd, s, size))
465 return FALSE;
467 data = bfd_zalloc (ibfd, size);
468 if (data == NULL)
469 return FALSE;
471 bfd_put_32 (ibfd, sizeof (SPU_PLUGIN_NAME), data + 0);
472 bfd_put_32 (ibfd, name_len, data + 4);
473 bfd_put_32 (ibfd, 1, data + 8);
474 memcpy (data + 12, SPU_PLUGIN_NAME, sizeof (SPU_PLUGIN_NAME));
475 memcpy (data + 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4),
476 bfd_get_filename (output_bfd), name_len);
477 s->contents = data;
480 return TRUE;
483 /* qsort predicate to sort sections by vma. */
485 static int
486 sort_sections (const void *a, const void *b)
488 const asection *const *s1 = a;
489 const asection *const *s2 = b;
490 bfd_signed_vma delta = (*s1)->vma - (*s2)->vma;
492 if (delta != 0)
493 return delta < 0 ? -1 : 1;
495 return (*s1)->index - (*s2)->index;
498 /* Identify overlays in the output bfd, and number them. */
500 bfd_boolean
501 spu_elf_find_overlays (bfd *output_bfd, struct bfd_link_info *info)
503 struct spu_link_hash_table *htab = spu_hash_table (info);
504 asection **alloc_sec;
505 unsigned int i, n, ovl_index, num_buf;
506 asection *s;
507 bfd_vma ovl_end;
509 if (output_bfd->section_count < 2)
510 return FALSE;
512 alloc_sec = bfd_malloc (output_bfd->section_count * sizeof (*alloc_sec));
513 if (alloc_sec == NULL)
514 return FALSE;
516 /* Pick out all the alloced sections. */
517 for (n = 0, s = output_bfd->sections; s != NULL; s = s->next)
518 if ((s->flags & SEC_ALLOC) != 0
519 && (s->flags & (SEC_LOAD | SEC_THREAD_LOCAL)) != SEC_THREAD_LOCAL
520 && s->size != 0)
521 alloc_sec[n++] = s;
523 if (n == 0)
525 free (alloc_sec);
526 return FALSE;
529 /* Sort them by vma. */
530 qsort (alloc_sec, n, sizeof (*alloc_sec), sort_sections);
532 /* Look for overlapping vmas. Any with overlap must be overlays.
533 Count them. Also count the number of overlay regions. */
534 ovl_end = alloc_sec[0]->vma + alloc_sec[0]->size;
535 for (ovl_index = 0, num_buf = 0, i = 1; i < n; i++)
537 s = alloc_sec[i];
538 if (s->vma < ovl_end)
540 asection *s0 = alloc_sec[i - 1];
542 if (spu_elf_section_data (s0)->u.o.ovl_index == 0)
544 alloc_sec[ovl_index] = s0;
545 spu_elf_section_data (s0)->u.o.ovl_index = ++ovl_index;
546 spu_elf_section_data (s0)->u.o.ovl_buf = ++num_buf;
548 alloc_sec[ovl_index] = s;
549 spu_elf_section_data (s)->u.o.ovl_index = ++ovl_index;
550 spu_elf_section_data (s)->u.o.ovl_buf = num_buf;
551 if (s0->vma != s->vma)
553 info->callbacks->einfo (_("%X%P: overlay sections %A and %A "
554 "do not start at the same address.\n"),
555 s0, s);
556 return FALSE;
558 if (ovl_end < s->vma + s->size)
559 ovl_end = s->vma + s->size;
561 else
562 ovl_end = s->vma + s->size;
565 htab->num_overlays = ovl_index;
566 htab->num_buf = num_buf;
567 htab->ovl_sec = alloc_sec;
568 return ovl_index != 0;
571 /* Support two sizes of overlay stubs, a slower more compact stub of two
572 intructions, and a faster stub of four instructions. */
573 #ifndef OVL_STUB_SIZE
574 /* Default to faster. */
575 #define OVL_STUB_SIZE 16
576 /* #define OVL_STUB_SIZE 8 */
577 #endif
578 #define BRSL 0x33000000
579 #define BR 0x32000000
580 #define NOP 0x40200000
581 #define LNOP 0x00200000
582 #define ILA 0x42000000
584 /* Return true for all relative and absolute branch instructions.
585 bra 00110000 0..
586 brasl 00110001 0..
587 br 00110010 0..
588 brsl 00110011 0..
589 brz 00100000 0..
590 brnz 00100001 0..
591 brhz 00100010 0..
592 brhnz 00100011 0.. */
594 static bfd_boolean
595 is_branch (const unsigned char *insn)
597 return (insn[0] & 0xec) == 0x20 && (insn[1] & 0x80) == 0;
600 /* Return true for all indirect branch instructions.
601 bi 00110101 000
602 bisl 00110101 001
603 iret 00110101 010
604 bisled 00110101 011
605 biz 00100101 000
606 binz 00100101 001
607 bihz 00100101 010
608 bihnz 00100101 011 */
610 static bfd_boolean
611 is_indirect_branch (const unsigned char *insn)
613 return (insn[0] & 0xef) == 0x25 && (insn[1] & 0x80) == 0;
616 /* Return true for branch hint instructions.
617 hbra 0001000..
618 hbrr 0001001.. */
620 static bfd_boolean
621 is_hint (const unsigned char *insn)
623 return (insn[0] & 0xfc) == 0x10;
626 /* Return TRUE if this reloc symbol should possibly go via an overlay stub. */
628 static bfd_boolean
629 needs_ovl_stub (const char *sym_name,
630 asection *sym_sec,
631 asection *input_section,
632 struct spu_link_hash_table *htab,
633 bfd_boolean is_branch)
635 if (htab->num_overlays == 0)
636 return FALSE;
638 if (sym_sec == NULL
639 || sym_sec->output_section == NULL
640 || spu_elf_section_data (sym_sec->output_section) == NULL)
641 return FALSE;
643 /* setjmp always goes via an overlay stub, because then the return
644 and hence the longjmp goes via __ovly_return. That magically
645 makes setjmp/longjmp between overlays work. */
646 if (strncmp (sym_name, "setjmp", 6) == 0
647 && (sym_name[6] == '\0' || sym_name[6] == '@'))
648 return TRUE;
650 /* Usually, symbols in non-overlay sections don't need stubs. */
651 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index == 0
652 && !htab->non_overlay_stubs)
653 return FALSE;
655 /* A reference from some other section to a symbol in an overlay
656 section needs a stub. */
657 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index
658 != spu_elf_section_data (input_section->output_section)->u.o.ovl_index)
659 return TRUE;
661 /* If this insn isn't a branch then we are possibly taking the
662 address of a function and passing it out somehow. */
663 return !is_branch;
666 enum _insn_type { non_branch, branch, call };
668 static bfd_boolean
669 count_stub (struct spu_link_hash_table *htab,
670 bfd *ibfd,
671 asection *isec,
672 enum _insn_type insn_type,
673 struct elf_link_hash_entry *h,
674 const Elf_Internal_Rela *irela)
676 unsigned int ovl = 0;
677 struct got_entry *g, **head;
679 /* If this instruction is a branch or call, we need a stub
680 for it. One stub per function per overlay.
681 If it isn't a branch, then we are taking the address of
682 this function so need a stub in the non-overlay area
683 for it. One stub per function. */
684 if (insn_type != non_branch)
685 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
687 if (h != NULL)
688 head = &h->got.glist;
689 else
691 if (elf_local_got_ents (ibfd) == NULL)
693 bfd_size_type amt = (elf_tdata (ibfd)->symtab_hdr.sh_info
694 * sizeof (*elf_local_got_ents (ibfd)));
695 elf_local_got_ents (ibfd) = bfd_zmalloc (amt);
696 if (elf_local_got_ents (ibfd) == NULL)
697 return FALSE;
699 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
702 /* If we have a stub in the non-overlay area then there's no need
703 for one in overlays. */
704 g = *head;
705 if (g != NULL && g->ovl == 0)
706 return TRUE;
708 if (ovl == 0)
710 struct got_entry *gnext;
712 /* Need a new non-overlay area stub. Zap other stubs. */
713 for (; g != NULL; g = gnext)
715 htab->stub_count[g->ovl] -= 1;
716 gnext = g->next;
717 free (g);
720 else
722 for (; g != NULL; g = g->next)
723 if (g->ovl == ovl)
724 break;
727 if (g == NULL)
729 g = bfd_malloc (sizeof *g);
730 if (g == NULL)
731 return FALSE;
732 g->ovl = ovl;
733 g->stub_addr = (bfd_vma) -1;
734 g->next = *head;
735 *head = g;
737 htab->stub_count[ovl] += 1;
740 return TRUE;
743 /* Two instruction overlay stubs look like:
745 brsl $75,__ovly_load
746 .word target_ovl_and_address
748 ovl_and_address is a word with the overlay number in the top 14 bits
749 and local store address in the bottom 18 bits.
751 Four instruction overlay stubs look like:
753 ila $78,ovl_number
754 lnop
755 ila $79,target_address
756 br __ovly_load */
758 static bfd_boolean
759 build_stub (struct spu_link_hash_table *htab,
760 bfd *ibfd,
761 asection *isec,
762 enum _insn_type insn_type,
763 struct elf_link_hash_entry *h,
764 const Elf_Internal_Rela *irela,
765 bfd_vma dest,
766 asection *dest_sec)
768 unsigned int ovl;
769 struct got_entry *g, **head;
770 asection *sec;
771 bfd_vma val, from, to;
773 ovl = 0;
774 if (insn_type != non_branch)
775 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
777 if (h != NULL)
778 head = &h->got.glist;
779 else
780 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
782 g = *head;
783 if (g != NULL && g->ovl == 0 && ovl != 0)
784 return TRUE;
786 for (; g != NULL; g = g->next)
787 if (g->ovl == ovl)
788 break;
789 if (g == NULL)
790 abort ();
792 if (g->stub_addr != (bfd_vma) -1)
793 return TRUE;
795 sec = htab->stub_sec[ovl];
796 dest += dest_sec->output_offset + dest_sec->output_section->vma;
797 from = sec->size + sec->output_offset + sec->output_section->vma;
798 g->stub_addr = from;
799 to = (htab->ovly_load->root.u.def.value
800 + htab->ovly_load->root.u.def.section->output_offset
801 + htab->ovly_load->root.u.def.section->output_section->vma);
802 val = to - from;
803 if (OVL_STUB_SIZE == 16)
804 val -= 12;
805 if (((dest | to | from) & 3) != 0
806 || val + 0x20000 >= 0x40000)
808 htab->stub_err = 1;
809 return FALSE;
811 ovl = spu_elf_section_data (dest_sec->output_section)->u.o.ovl_index;
813 if (OVL_STUB_SIZE == 16)
815 bfd_put_32 (sec->owner, ILA + ((ovl << 7) & 0x01ffff80) + 78,
816 sec->contents + sec->size);
817 bfd_put_32 (sec->owner, LNOP,
818 sec->contents + sec->size + 4);
819 bfd_put_32 (sec->owner, ILA + ((dest << 7) & 0x01ffff80) + 79,
820 sec->contents + sec->size + 8);
821 bfd_put_32 (sec->owner, BR + ((val << 5) & 0x007fff80),
822 sec->contents + sec->size + 12);
824 else if (OVL_STUB_SIZE == 8)
826 bfd_put_32 (sec->owner, BRSL + ((val << 5) & 0x007fff80) + 75,
827 sec->contents + sec->size);
829 val = (dest & 0x3ffff) | (ovl << 14);
830 bfd_put_32 (sec->owner, val,
831 sec->contents + sec->size + 4);
833 else
834 abort ();
835 sec->size += OVL_STUB_SIZE;
837 if (htab->emit_stub_syms)
839 size_t len;
840 char *name;
841 int add;
843 len = 8 + sizeof (".ovl_call.") - 1;
844 if (h != NULL)
845 len += strlen (h->root.root.string);
846 else
847 len += 8 + 1 + 8;
848 add = 0;
849 if (irela != NULL)
850 add = (int) irela->r_addend & 0xffffffff;
851 if (add != 0)
852 len += 1 + 8;
853 name = bfd_malloc (len);
854 if (name == NULL)
855 return FALSE;
857 sprintf (name, "%08x.ovl_call.", g->ovl);
858 if (h != NULL)
859 strcpy (name + 8 + sizeof (".ovl_call.") - 1, h->root.root.string);
860 else
861 sprintf (name + 8 + sizeof (".ovl_call.") - 1, "%x:%x",
862 dest_sec->id & 0xffffffff,
863 (int) ELF32_R_SYM (irela->r_info) & 0xffffffff);
864 if (add != 0)
865 sprintf (name + len - 9, "+%x", add);
867 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
868 free (name);
869 if (h == NULL)
870 return FALSE;
871 if (h->root.type == bfd_link_hash_new)
873 h->root.type = bfd_link_hash_defined;
874 h->root.u.def.section = sec;
875 h->root.u.def.value = sec->size - OVL_STUB_SIZE;
876 h->size = OVL_STUB_SIZE;
877 h->type = STT_FUNC;
878 h->ref_regular = 1;
879 h->def_regular = 1;
880 h->ref_regular_nonweak = 1;
881 h->forced_local = 1;
882 h->non_elf = 0;
886 return TRUE;
889 /* Called via elf_link_hash_traverse to allocate stubs for any _SPUEAR_
890 symbols. */
892 static bfd_boolean
893 allocate_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
895 /* Symbols starting with _SPUEAR_ need a stub because they may be
896 invoked by the PPU. */
897 if ((h->root.type == bfd_link_hash_defined
898 || h->root.type == bfd_link_hash_defweak)
899 && h->def_regular
900 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0)
902 struct spu_link_hash_table *htab = inf;
904 count_stub (htab, NULL, NULL, non_branch, h, NULL);
907 return TRUE;
910 static bfd_boolean
911 build_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
913 /* Symbols starting with _SPUEAR_ need a stub because they may be
914 invoked by the PPU. */
915 if ((h->root.type == bfd_link_hash_defined
916 || h->root.type == bfd_link_hash_defweak)
917 && h->def_regular
918 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0)
920 struct spu_link_hash_table *htab = inf;
922 build_stub (htab, NULL, NULL, non_branch, h, NULL,
923 h->root.u.def.value, h->root.u.def.section);
926 return TRUE;
929 /* Size or build stubs. */
931 static bfd_boolean
932 process_stubs (bfd *output_bfd,
933 struct bfd_link_info *info,
934 bfd_boolean build)
936 struct spu_link_hash_table *htab = spu_hash_table (info);
937 bfd *ibfd;
939 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
941 extern const bfd_target bfd_elf32_spu_vec;
942 Elf_Internal_Shdr *symtab_hdr;
943 asection *isec;
944 Elf_Internal_Sym *local_syms = NULL;
945 void *psyms;
947 if (ibfd->xvec != &bfd_elf32_spu_vec)
948 continue;
950 /* We'll need the symbol table in a second. */
951 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
952 if (symtab_hdr->sh_info == 0)
953 continue;
955 /* Arrange to read and keep global syms for later stack analysis. */
956 psyms = &local_syms;
957 if (htab->stack_analysis)
958 psyms = &symtab_hdr->contents;
960 /* Walk over each section attached to the input bfd. */
961 for (isec = ibfd->sections; isec != NULL; isec = isec->next)
963 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
965 /* If there aren't any relocs, then there's nothing more to do. */
966 if ((isec->flags & SEC_RELOC) == 0
967 || (isec->flags & SEC_ALLOC) == 0
968 || (isec->flags & SEC_LOAD) == 0
969 || isec->reloc_count == 0)
970 continue;
972 /* If this section is a link-once section that will be
973 discarded, then don't create any stubs. */
974 if (isec->output_section == NULL
975 || isec->output_section->owner != output_bfd)
976 continue;
978 /* Get the relocs. */
979 internal_relocs = _bfd_elf_link_read_relocs (ibfd, isec, NULL, NULL,
980 info->keep_memory);
981 if (internal_relocs == NULL)
982 goto error_ret_free_local;
984 /* Now examine each relocation. */
985 irela = internal_relocs;
986 irelaend = irela + isec->reloc_count;
987 for (; irela < irelaend; irela++)
989 enum elf_spu_reloc_type r_type;
990 unsigned int r_indx;
991 asection *sym_sec;
992 Elf_Internal_Sym *sym;
993 struct elf_link_hash_entry *h;
994 const char *sym_name;
995 unsigned int sym_type;
996 enum _insn_type insn_type;
998 r_type = ELF32_R_TYPE (irela->r_info);
999 r_indx = ELF32_R_SYM (irela->r_info);
1001 if (r_type >= R_SPU_max)
1003 bfd_set_error (bfd_error_bad_value);
1004 error_ret_free_internal:
1005 if (elf_section_data (isec)->relocs != internal_relocs)
1006 free (internal_relocs);
1007 error_ret_free_local:
1008 if (local_syms != NULL
1009 && (symtab_hdr->contents
1010 != (unsigned char *) local_syms))
1011 free (local_syms);
1012 return FALSE;
1015 /* Determine the reloc target section. */
1016 if (!get_sym_h (&h, &sym, &sym_sec, psyms, r_indx, ibfd))
1017 goto error_ret_free_internal;
1019 if (sym_sec == NULL
1020 || sym_sec->output_section == NULL
1021 || sym_sec->output_section->owner != output_bfd)
1022 continue;
1024 /* Ensure no stubs for user supplied overlay manager syms. */
1025 if (h != NULL
1026 && (strcmp (h->root.root.string, "__ovly_load") == 0
1027 || strcmp (h->root.root.string, "__ovly_return") == 0))
1028 continue;
1030 insn_type = non_branch;
1031 if (r_type == R_SPU_REL16
1032 || r_type == R_SPU_ADDR16)
1034 unsigned char insn[4];
1036 if (!bfd_get_section_contents (ibfd, isec, insn,
1037 irela->r_offset, 4))
1038 goto error_ret_free_internal;
1040 if (is_branch (insn) || is_hint (insn))
1042 insn_type = branch;
1043 if ((insn[0] & 0xfd) == 0x31)
1044 insn_type = call;
1048 /* We are only interested in function symbols. */
1049 if (h != NULL)
1051 sym_type = h->type;
1052 sym_name = h->root.root.string;
1054 else
1056 sym_type = ELF_ST_TYPE (sym->st_info);
1057 sym_name = bfd_elf_sym_name (sym_sec->owner,
1058 symtab_hdr,
1059 sym,
1060 sym_sec);
1063 if (sym_type != STT_FUNC)
1065 /* It's common for people to write assembly and forget
1066 to give function symbols the right type. Handle
1067 calls to such symbols, but warn so that (hopefully)
1068 people will fix their code. We need the symbol
1069 type to be correct to distinguish function pointer
1070 initialisation from other pointer initialisation. */
1071 if (insn_type == call)
1072 (*_bfd_error_handler) (_("warning: call to non-function"
1073 " symbol %s defined in %B"),
1074 sym_sec->owner, sym_name);
1075 else if (insn_type == non_branch)
1076 continue;
1079 if (!needs_ovl_stub (sym_name, sym_sec, isec, htab,
1080 insn_type != non_branch))
1081 continue;
1083 if (htab->stub_count == NULL)
1085 bfd_size_type amt;
1086 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_count);
1087 htab->stub_count = bfd_zmalloc (amt);
1088 if (htab->stub_count == NULL)
1089 goto error_ret_free_internal;
1092 if (!build)
1094 if (!count_stub (htab, ibfd, isec, insn_type, h, irela))
1095 goto error_ret_free_internal;
1097 else
1099 bfd_vma dest;
1101 if (h != NULL)
1102 dest = h->root.u.def.value;
1103 else
1104 dest = sym->st_value;
1105 if (!build_stub (htab, ibfd, isec, insn_type, h, irela,
1106 dest, sym_sec))
1107 goto error_ret_free_internal;
1111 /* We're done with the internal relocs, free them. */
1112 if (elf_section_data (isec)->relocs != internal_relocs)
1113 free (internal_relocs);
1116 if (local_syms != NULL
1117 && symtab_hdr->contents != (unsigned char *) local_syms)
1119 if (!info->keep_memory)
1120 free (local_syms);
1121 else
1122 symtab_hdr->contents = (unsigned char *) local_syms;
1126 return TRUE;
1129 /* Allocate space for overlay call and return stubs. */
1132 spu_elf_size_stubs (bfd *output_bfd,
1133 struct bfd_link_info *info,
1134 void (*place_spu_section) (asection *, asection *,
1135 const char *),
1136 int non_overlay_stubs)
1138 struct spu_link_hash_table *htab = spu_hash_table (info);
1139 bfd *ibfd;
1140 bfd_size_type amt;
1141 flagword flags;
1142 unsigned int i;
1143 asection *stub;
1145 htab->non_overlay_stubs = non_overlay_stubs;
1146 if (!process_stubs (output_bfd, info, FALSE))
1147 return 0;
1149 elf_link_hash_traverse (&htab->elf, allocate_spuear_stubs, htab);
1150 if (htab->stub_err)
1151 return 0;
1153 if (htab->stub_count == NULL)
1154 return 1;
1156 ibfd = info->input_bfds;
1157 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_sec);
1158 htab->stub_sec = bfd_zmalloc (amt);
1159 if (htab->stub_sec == NULL)
1160 return 0;
1162 flags = (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_READONLY
1163 | SEC_HAS_CONTENTS | SEC_IN_MEMORY);
1164 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1165 htab->stub_sec[0] = stub;
1166 if (stub == NULL
1167 || !bfd_set_section_alignment (ibfd, stub, 3 + (OVL_STUB_SIZE > 8)))
1168 return 0;
1169 stub->size = htab->stub_count[0] * OVL_STUB_SIZE;
1170 (*place_spu_section) (stub, NULL, ".text");
1172 for (i = 0; i < htab->num_overlays; ++i)
1174 asection *osec = htab->ovl_sec[i];
1175 unsigned int ovl = spu_elf_section_data (osec)->u.o.ovl_index;
1176 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1177 htab->stub_sec[ovl] = stub;
1178 if (stub == NULL
1179 || !bfd_set_section_alignment (ibfd, stub, 3 + (OVL_STUB_SIZE > 8)))
1180 return 0;
1181 stub->size = htab->stub_count[ovl] * OVL_STUB_SIZE;
1182 (*place_spu_section) (stub, osec, NULL);
1185 /* htab->ovtab consists of two arrays.
1186 . struct {
1187 . u32 vma;
1188 . u32 size;
1189 . u32 file_off;
1190 . u32 buf;
1191 . } _ovly_table[];
1193 . struct {
1194 . u32 mapped;
1195 . } _ovly_buf_table[];
1196 . */
1198 flags = (SEC_ALLOC | SEC_LOAD
1199 | SEC_HAS_CONTENTS | SEC_IN_MEMORY);
1200 htab->ovtab = bfd_make_section_anyway_with_flags (ibfd, ".ovtab", flags);
1201 if (htab->ovtab == NULL
1202 || !bfd_set_section_alignment (ibfd, htab->ovtab, 4))
1203 return 0;
1205 htab->ovtab->size = htab->num_overlays * 16 + 16 + htab->num_buf * 2 * 4;
1206 (*place_spu_section) (htab->ovtab, NULL, ".data");
1208 htab->toe = bfd_make_section_anyway_with_flags (ibfd, ".toe", SEC_ALLOC);
1209 if (htab->toe == NULL
1210 || !bfd_set_section_alignment (ibfd, htab->toe, 4))
1211 return 0;
1212 htab->toe->size = 16;
1213 (*place_spu_section) (htab->toe, NULL, ".toe");
1215 return 2;
1218 /* Functions to handle embedded spu_ovl.o object. */
1220 static void *
1221 ovl_mgr_open (struct bfd *nbfd ATTRIBUTE_UNUSED, void *stream)
1223 return stream;
1226 static file_ptr
1227 ovl_mgr_pread (struct bfd *abfd ATTRIBUTE_UNUSED,
1228 void *stream,
1229 void *buf,
1230 file_ptr nbytes,
1231 file_ptr offset)
1233 struct _ovl_stream *os;
1234 size_t count;
1235 size_t max;
1237 os = (struct _ovl_stream *) stream;
1238 max = (const char *) os->end - (const char *) os->start;
1240 if ((ufile_ptr) offset >= max)
1241 return 0;
1243 count = nbytes;
1244 if (count > max - offset)
1245 count = max - offset;
1247 memcpy (buf, (const char *) os->start + offset, count);
1248 return count;
1251 bfd_boolean
1252 spu_elf_open_builtin_lib (bfd **ovl_bfd, const struct _ovl_stream *stream)
1254 *ovl_bfd = bfd_openr_iovec ("builtin ovl_mgr",
1255 "elf32-spu",
1256 ovl_mgr_open,
1257 (void *) stream,
1258 ovl_mgr_pread,
1259 NULL,
1260 NULL);
1261 return *ovl_bfd != NULL;
1264 /* Define an STT_OBJECT symbol. */
1266 static struct elf_link_hash_entry *
1267 define_ovtab_symbol (struct spu_link_hash_table *htab, const char *name)
1269 struct elf_link_hash_entry *h;
1271 h = elf_link_hash_lookup (&htab->elf, name, TRUE, FALSE, FALSE);
1272 if (h == NULL)
1273 return NULL;
1275 if (h->root.type != bfd_link_hash_defined
1276 || !h->def_regular)
1278 h->root.type = bfd_link_hash_defined;
1279 h->root.u.def.section = htab->ovtab;
1280 h->type = STT_OBJECT;
1281 h->ref_regular = 1;
1282 h->def_regular = 1;
1283 h->ref_regular_nonweak = 1;
1284 h->non_elf = 0;
1286 else
1288 (*_bfd_error_handler) (_("%B is not allowed to define %s"),
1289 h->root.u.def.section->owner,
1290 h->root.root.string);
1291 bfd_set_error (bfd_error_bad_value);
1292 return NULL;
1295 return h;
1298 /* Fill in all stubs and the overlay tables. */
1300 bfd_boolean
1301 spu_elf_build_stubs (struct bfd_link_info *info, int emit_syms)
1303 struct spu_link_hash_table *htab = spu_hash_table (info);
1304 struct elf_link_hash_entry *h;
1305 bfd_byte *p;
1306 asection *s;
1307 bfd *obfd;
1308 unsigned int i;
1310 htab->emit_stub_syms = emit_syms;
1311 if (htab->stub_count == NULL)
1312 return TRUE;
1314 for (i = 0; i <= htab->num_overlays; i++)
1315 if (htab->stub_sec[i]->size != 0)
1317 htab->stub_sec[i]->contents = bfd_zalloc (htab->stub_sec[i]->owner,
1318 htab->stub_sec[i]->size);
1319 if (htab->stub_sec[i]->contents == NULL)
1320 return FALSE;
1321 htab->stub_sec[i]->rawsize = htab->stub_sec[i]->size;
1322 htab->stub_sec[i]->size = 0;
1325 h = elf_link_hash_lookup (&htab->elf, "__ovly_load", FALSE, FALSE, FALSE);
1326 htab->ovly_load = h;
1327 BFD_ASSERT (h != NULL
1328 && (h->root.type == bfd_link_hash_defined
1329 || h->root.type == bfd_link_hash_defweak)
1330 && h->def_regular);
1332 s = h->root.u.def.section->output_section;
1333 if (spu_elf_section_data (s)->u.o.ovl_index)
1335 (*_bfd_error_handler) (_("%s in overlay section"),
1336 h->root.u.def.section->owner);
1337 bfd_set_error (bfd_error_bad_value);
1338 return FALSE;
1341 h = elf_link_hash_lookup (&htab->elf, "__ovly_return", FALSE, FALSE, FALSE);
1342 htab->ovly_return = h;
1344 /* Write out all the stubs. */
1345 obfd = htab->ovtab->output_section->owner;
1346 process_stubs (obfd, info, TRUE);
1348 elf_link_hash_traverse (&htab->elf, build_spuear_stubs, htab);
1349 if (htab->stub_err)
1350 return FALSE;
1352 for (i = 0; i <= htab->num_overlays; i++)
1354 if (htab->stub_sec[i]->size != htab->stub_sec[i]->rawsize)
1356 (*_bfd_error_handler) (_("stubs don't match calculated size"));
1357 bfd_set_error (bfd_error_bad_value);
1358 return FALSE;
1360 htab->stub_sec[i]->rawsize = 0;
1363 if (htab->stub_err)
1365 (*_bfd_error_handler) (_("overlay stub relocation overflow"));
1366 bfd_set_error (bfd_error_bad_value);
1367 return FALSE;
1370 htab->ovtab->contents = bfd_zalloc (htab->ovtab->owner, htab->ovtab->size);
1371 if (htab->ovtab->contents == NULL)
1372 return FALSE;
1374 /* Write out _ovly_table. */
1375 p = htab->ovtab->contents;
1376 /* set low bit of .buf to mark non-overlay area as present. */
1377 p[15] = 1;
1378 for (s = obfd->sections; s != NULL; s = s->next)
1380 unsigned int ovl_index = spu_elf_section_data (s)->u.o.ovl_index;
1382 if (ovl_index != 0)
1384 unsigned long off = ovl_index * 16;
1385 unsigned int ovl_buf = spu_elf_section_data (s)->u.o.ovl_buf;
1387 bfd_put_32 (htab->ovtab->owner, s->vma, p + off);
1388 bfd_put_32 (htab->ovtab->owner, (s->size + 15) & -16, p + off + 4);
1389 /* file_off written later in spu_elf_modify_program_headers. */
1390 bfd_put_32 (htab->ovtab->owner, ovl_buf * 2, p + off + 12);
1394 h = define_ovtab_symbol (htab, "_ovly_table");
1395 if (h == NULL)
1396 return FALSE;
1397 h->root.u.def.value = 16;
1398 h->size = htab->num_overlays * 16;
1400 h = define_ovtab_symbol (htab, "_ovly_table_end");
1401 if (h == NULL)
1402 return FALSE;
1403 h->root.u.def.value = htab->num_overlays * 16 + 16;
1404 h->size = 0;
1406 h = define_ovtab_symbol (htab, "_ovly_buf_table");
1407 if (h == NULL)
1408 return FALSE;
1409 h->root.u.def.value = htab->num_overlays * 16 + 16;
1410 h->size = htab->num_buf * 2 * 4;
1412 h = define_ovtab_symbol (htab, "_ovly_buf_table_end");
1413 if (h == NULL)
1414 return FALSE;
1415 h->root.u.def.value = htab->num_overlays * 16 + 16 + htab->num_buf * 2 * 4;
1416 h->size = 0;
1418 h = define_ovtab_symbol (htab, "_EAR_");
1419 if (h == NULL)
1420 return FALSE;
1421 h->root.u.def.section = htab->toe;
1422 h->root.u.def.value = 0;
1423 h->size = 16;
1425 return TRUE;
1428 /* OFFSET in SEC (presumably) is the beginning of a function prologue.
1429 Search for stack adjusting insns, and return the sp delta. */
1431 static int
1432 find_function_stack_adjust (asection *sec, bfd_vma offset)
1434 int unrecog;
1435 int reg[128];
1437 memset (reg, 0, sizeof (reg));
1438 for (unrecog = 0; offset + 4 <= sec->size && unrecog < 32; offset += 4)
1440 unsigned char buf[4];
1441 int rt, ra;
1442 int imm;
1444 /* Assume no relocs on stack adjusing insns. */
1445 if (!bfd_get_section_contents (sec->owner, sec, buf, offset, 4))
1446 break;
1448 if (buf[0] == 0x24 /* stqd */)
1449 continue;
1451 rt = buf[3] & 0x7f;
1452 ra = ((buf[2] & 0x3f) << 1) | (buf[3] >> 7);
1453 /* Partly decoded immediate field. */
1454 imm = (buf[1] << 9) | (buf[2] << 1) | (buf[3] >> 7);
1456 if (buf[0] == 0x1c /* ai */)
1458 imm >>= 7;
1459 imm = (imm ^ 0x200) - 0x200;
1460 reg[rt] = reg[ra] + imm;
1462 if (rt == 1 /* sp */)
1464 if (imm > 0)
1465 break;
1466 return reg[rt];
1469 else if (buf[0] == 0x18 && (buf[1] & 0xe0) == 0 /* a */)
1471 int rb = ((buf[1] & 0x1f) << 2) | ((buf[2] & 0xc0) >> 6);
1473 reg[rt] = reg[ra] + reg[rb];
1474 if (rt == 1)
1475 return reg[rt];
1477 else if ((buf[0] & 0xfc) == 0x40 /* il, ilh, ilhu, ila */)
1479 if (buf[0] >= 0x42 /* ila */)
1480 imm |= (buf[0] & 1) << 17;
1481 else
1483 imm &= 0xffff;
1485 if (buf[0] == 0x40 /* il */)
1487 if ((buf[1] & 0x80) == 0)
1488 goto unknown_insn;
1489 imm = (imm ^ 0x8000) - 0x8000;
1491 else if ((buf[1] & 0x80) == 0 /* ilhu */)
1492 imm <<= 16;
1494 reg[rt] = imm;
1495 continue;
1497 else if (buf[0] == 0x60 && (buf[1] & 0x80) != 0 /* iohl */)
1499 reg[rt] |= imm & 0xffff;
1500 continue;
1502 else if (buf[0] == 0x04 /* ori */)
1504 imm >>= 7;
1505 imm = (imm ^ 0x200) - 0x200;
1506 reg[rt] = reg[ra] | imm;
1507 continue;
1509 else if ((buf[0] == 0x33 && imm == 1 /* brsl .+4 */)
1510 || (buf[0] == 0x08 && (buf[1] & 0xe0) == 0 /* sf */))
1512 /* Used in pic reg load. Say rt is trashed. */
1513 reg[rt] = 0;
1514 continue;
1516 else if (is_branch (buf) || is_indirect_branch (buf))
1517 /* If we hit a branch then we must be out of the prologue. */
1518 break;
1519 unknown_insn:
1520 ++unrecog;
1523 return 0;
1526 /* qsort predicate to sort symbols by section and value. */
1528 static Elf_Internal_Sym *sort_syms_syms;
1529 static asection **sort_syms_psecs;
1531 static int
1532 sort_syms (const void *a, const void *b)
1534 Elf_Internal_Sym *const *s1 = a;
1535 Elf_Internal_Sym *const *s2 = b;
1536 asection *sec1,*sec2;
1537 bfd_signed_vma delta;
1539 sec1 = sort_syms_psecs[*s1 - sort_syms_syms];
1540 sec2 = sort_syms_psecs[*s2 - sort_syms_syms];
1542 if (sec1 != sec2)
1543 return sec1->index - sec2->index;
1545 delta = (*s1)->st_value - (*s2)->st_value;
1546 if (delta != 0)
1547 return delta < 0 ? -1 : 1;
1549 delta = (*s2)->st_size - (*s1)->st_size;
1550 if (delta != 0)
1551 return delta < 0 ? -1 : 1;
1553 return *s1 < *s2 ? -1 : 1;
1556 struct call_info
1558 struct function_info *fun;
1559 struct call_info *next;
1560 int is_tail;
1563 struct function_info
1565 /* List of functions called. Also branches to hot/cold part of
1566 function. */
1567 struct call_info *call_list;
1568 /* For hot/cold part of function, point to owner. */
1569 struct function_info *start;
1570 /* Symbol at start of function. */
1571 union {
1572 Elf_Internal_Sym *sym;
1573 struct elf_link_hash_entry *h;
1574 } u;
1575 /* Function section. */
1576 asection *sec;
1577 /* Address range of (this part of) function. */
1578 bfd_vma lo, hi;
1579 /* Stack usage. */
1580 int stack;
1581 /* Set if global symbol. */
1582 unsigned int global : 1;
1583 /* Set if known to be start of function (as distinct from a hunk
1584 in hot/cold section. */
1585 unsigned int is_func : 1;
1586 /* Flags used during call tree traversal. */
1587 unsigned int visit1 : 1;
1588 unsigned int non_root : 1;
1589 unsigned int visit2 : 1;
1590 unsigned int marking : 1;
1591 unsigned int visit3 : 1;
1594 struct spu_elf_stack_info
1596 int num_fun;
1597 int max_fun;
1598 /* Variable size array describing functions, one per contiguous
1599 address range belonging to a function. */
1600 struct function_info fun[1];
1603 /* Allocate a struct spu_elf_stack_info with MAX_FUN struct function_info
1604 entries for section SEC. */
1606 static struct spu_elf_stack_info *
1607 alloc_stack_info (asection *sec, int max_fun)
1609 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
1610 bfd_size_type amt;
1612 amt = sizeof (struct spu_elf_stack_info);
1613 amt += (max_fun - 1) * sizeof (struct function_info);
1614 sec_data->u.i.stack_info = bfd_zmalloc (amt);
1615 if (sec_data->u.i.stack_info != NULL)
1616 sec_data->u.i.stack_info->max_fun = max_fun;
1617 return sec_data->u.i.stack_info;
1620 /* Add a new struct function_info describing a (part of a) function
1621 starting at SYM_H. Keep the array sorted by address. */
1623 static struct function_info *
1624 maybe_insert_function (asection *sec,
1625 void *sym_h,
1626 bfd_boolean global,
1627 bfd_boolean is_func)
1629 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
1630 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
1631 int i;
1632 bfd_vma off, size;
1634 if (sinfo == NULL)
1636 sinfo = alloc_stack_info (sec, 20);
1637 if (sinfo == NULL)
1638 return NULL;
1641 if (!global)
1643 Elf_Internal_Sym *sym = sym_h;
1644 off = sym->st_value;
1645 size = sym->st_size;
1647 else
1649 struct elf_link_hash_entry *h = sym_h;
1650 off = h->root.u.def.value;
1651 size = h->size;
1654 for (i = sinfo->num_fun; --i >= 0; )
1655 if (sinfo->fun[i].lo <= off)
1656 break;
1658 if (i >= 0)
1660 /* Don't add another entry for an alias, but do update some
1661 info. */
1662 if (sinfo->fun[i].lo == off)
1664 /* Prefer globals over local syms. */
1665 if (global && !sinfo->fun[i].global)
1667 sinfo->fun[i].global = TRUE;
1668 sinfo->fun[i].u.h = sym_h;
1670 if (is_func)
1671 sinfo->fun[i].is_func = TRUE;
1672 return &sinfo->fun[i];
1674 /* Ignore a zero-size symbol inside an existing function. */
1675 else if (sinfo->fun[i].hi > off && size == 0)
1676 return &sinfo->fun[i];
1679 if (++i < sinfo->num_fun)
1680 memmove (&sinfo->fun[i + 1], &sinfo->fun[i],
1681 (sinfo->num_fun - i) * sizeof (sinfo->fun[i]));
1682 else if (i >= sinfo->max_fun)
1684 bfd_size_type amt = sizeof (struct spu_elf_stack_info);
1685 bfd_size_type old = amt;
1687 old += (sinfo->max_fun - 1) * sizeof (struct function_info);
1688 sinfo->max_fun += 20 + (sinfo->max_fun >> 1);
1689 amt += (sinfo->max_fun - 1) * sizeof (struct function_info);
1690 sinfo = bfd_realloc (sinfo, amt);
1691 if (sinfo == NULL)
1692 return NULL;
1693 memset ((char *) sinfo + old, 0, amt - old);
1694 sec_data->u.i.stack_info = sinfo;
1696 sinfo->fun[i].is_func = is_func;
1697 sinfo->fun[i].global = global;
1698 sinfo->fun[i].sec = sec;
1699 if (global)
1700 sinfo->fun[i].u.h = sym_h;
1701 else
1702 sinfo->fun[i].u.sym = sym_h;
1703 sinfo->fun[i].lo = off;
1704 sinfo->fun[i].hi = off + size;
1705 sinfo->fun[i].stack = -find_function_stack_adjust (sec, off);
1706 sinfo->num_fun += 1;
1707 return &sinfo->fun[i];
1710 /* Return the name of FUN. */
1712 static const char *
1713 func_name (struct function_info *fun)
1715 asection *sec;
1716 bfd *ibfd;
1717 Elf_Internal_Shdr *symtab_hdr;
1719 while (fun->start != NULL)
1720 fun = fun->start;
1722 if (fun->global)
1723 return fun->u.h->root.root.string;
1725 sec = fun->sec;
1726 if (fun->u.sym->st_name == 0)
1728 size_t len = strlen (sec->name);
1729 char *name = bfd_malloc (len + 10);
1730 if (name == NULL)
1731 return "(null)";
1732 sprintf (name, "%s+%lx", sec->name,
1733 (unsigned long) fun->u.sym->st_value & 0xffffffff);
1734 return name;
1736 ibfd = sec->owner;
1737 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
1738 return bfd_elf_sym_name (ibfd, symtab_hdr, fun->u.sym, sec);
1741 /* Read the instruction at OFF in SEC. Return true iff the instruction
1742 is a nop, lnop, or stop 0 (all zero insn). */
1744 static bfd_boolean
1745 is_nop (asection *sec, bfd_vma off)
1747 unsigned char insn[4];
1749 if (off + 4 > sec->size
1750 || !bfd_get_section_contents (sec->owner, sec, insn, off, 4))
1751 return FALSE;
1752 if ((insn[0] & 0xbf) == 0 && (insn[1] & 0xe0) == 0x20)
1753 return TRUE;
1754 if (insn[0] == 0 && insn[1] == 0 && insn[2] == 0 && insn[3] == 0)
1755 return TRUE;
1756 return FALSE;
1759 /* Extend the range of FUN to cover nop padding up to LIMIT.
1760 Return TRUE iff some instruction other than a NOP was found. */
1762 static bfd_boolean
1763 insns_at_end (struct function_info *fun, bfd_vma limit)
1765 bfd_vma off = (fun->hi + 3) & -4;
1767 while (off < limit && is_nop (fun->sec, off))
1768 off += 4;
1769 if (off < limit)
1771 fun->hi = off;
1772 return TRUE;
1774 fun->hi = limit;
1775 return FALSE;
1778 /* Check and fix overlapping function ranges. Return TRUE iff there
1779 are gaps in the current info we have about functions in SEC. */
1781 static bfd_boolean
1782 check_function_ranges (asection *sec, struct bfd_link_info *info)
1784 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
1785 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
1786 int i;
1787 bfd_boolean gaps = FALSE;
1789 if (sinfo == NULL)
1790 return FALSE;
1792 for (i = 1; i < sinfo->num_fun; i++)
1793 if (sinfo->fun[i - 1].hi > sinfo->fun[i].lo)
1795 /* Fix overlapping symbols. */
1796 const char *f1 = func_name (&sinfo->fun[i - 1]);
1797 const char *f2 = func_name (&sinfo->fun[i]);
1799 info->callbacks->einfo (_("warning: %s overlaps %s\n"), f1, f2);
1800 sinfo->fun[i - 1].hi = sinfo->fun[i].lo;
1802 else if (insns_at_end (&sinfo->fun[i - 1], sinfo->fun[i].lo))
1803 gaps = TRUE;
1805 if (sinfo->num_fun == 0)
1806 gaps = TRUE;
1807 else
1809 if (sinfo->fun[0].lo != 0)
1810 gaps = TRUE;
1811 if (sinfo->fun[sinfo->num_fun - 1].hi > sec->size)
1813 const char *f1 = func_name (&sinfo->fun[sinfo->num_fun - 1]);
1815 info->callbacks->einfo (_("warning: %s exceeds section size\n"), f1);
1816 sinfo->fun[sinfo->num_fun - 1].hi = sec->size;
1818 else if (insns_at_end (&sinfo->fun[sinfo->num_fun - 1], sec->size))
1819 gaps = TRUE;
1821 return gaps;
1824 /* Search current function info for a function that contains address
1825 OFFSET in section SEC. */
1827 static struct function_info *
1828 find_function (asection *sec, bfd_vma offset, struct bfd_link_info *info)
1830 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
1831 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
1832 int lo, hi, mid;
1834 lo = 0;
1835 hi = sinfo->num_fun;
1836 while (lo < hi)
1838 mid = (lo + hi) / 2;
1839 if (offset < sinfo->fun[mid].lo)
1840 hi = mid;
1841 else if (offset >= sinfo->fun[mid].hi)
1842 lo = mid + 1;
1843 else
1844 return &sinfo->fun[mid];
1846 info->callbacks->einfo (_("%A:0x%v not found in function table\n"),
1847 sec, offset);
1848 return NULL;
1851 /* Add CALLEE to CALLER call list if not already present. */
1853 static bfd_boolean
1854 insert_callee (struct function_info *caller, struct call_info *callee)
1856 struct call_info *p;
1857 for (p = caller->call_list; p != NULL; p = p->next)
1858 if (p->fun == callee->fun)
1860 /* Tail calls use less stack than normal calls. Retain entry
1861 for normal call over one for tail call. */
1862 if (p->is_tail > callee->is_tail)
1863 p->is_tail = callee->is_tail;
1864 return FALSE;
1866 callee->next = caller->call_list;
1867 caller->call_list = callee;
1868 return TRUE;
1871 /* Rummage through the relocs for SEC, looking for function calls.
1872 If CALL_TREE is true, fill in call graph. If CALL_TREE is false,
1873 mark destination symbols on calls as being functions. Also
1874 look at branches, which may be tail calls or go to hot/cold
1875 section part of same function. */
1877 static bfd_boolean
1878 mark_functions_via_relocs (asection *sec,
1879 struct bfd_link_info *info,
1880 int call_tree)
1882 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
1883 Elf_Internal_Shdr *symtab_hdr = &elf_tdata (sec->owner)->symtab_hdr;
1884 Elf_Internal_Sym *syms;
1885 void *psyms;
1886 static bfd_boolean warned;
1888 internal_relocs = _bfd_elf_link_read_relocs (sec->owner, sec, NULL, NULL,
1889 info->keep_memory);
1890 if (internal_relocs == NULL)
1891 return FALSE;
1893 symtab_hdr = &elf_tdata (sec->owner)->symtab_hdr;
1894 psyms = &symtab_hdr->contents;
1895 syms = *(Elf_Internal_Sym **) psyms;
1896 irela = internal_relocs;
1897 irelaend = irela + sec->reloc_count;
1898 for (; irela < irelaend; irela++)
1900 enum elf_spu_reloc_type r_type;
1901 unsigned int r_indx;
1902 asection *sym_sec;
1903 Elf_Internal_Sym *sym;
1904 struct elf_link_hash_entry *h;
1905 bfd_vma val;
1906 unsigned char insn[4];
1907 bfd_boolean is_call;
1908 struct function_info *caller;
1909 struct call_info *callee;
1911 r_type = ELF32_R_TYPE (irela->r_info);
1912 if (r_type != R_SPU_REL16
1913 && r_type != R_SPU_ADDR16)
1914 continue;
1916 r_indx = ELF32_R_SYM (irela->r_info);
1917 if (!get_sym_h (&h, &sym, &sym_sec, psyms, r_indx, sec->owner))
1918 return FALSE;
1920 if (sym_sec == NULL
1921 || sym_sec->output_section == NULL
1922 || sym_sec->output_section->owner != sec->output_section->owner)
1923 continue;
1925 if (!bfd_get_section_contents (sec->owner, sec, insn,
1926 irela->r_offset, 4))
1927 return FALSE;
1928 if (!is_branch (insn))
1929 continue;
1931 if ((sym_sec->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE))
1932 != (SEC_ALLOC | SEC_LOAD | SEC_CODE))
1934 if (!call_tree)
1935 warned = TRUE;
1936 if (!call_tree || !warned)
1937 info->callbacks->einfo (_("%B(%A+0x%v): call to non-code section"
1938 " %B(%A), stack analysis incomplete\n"),
1939 sec->owner, sec, irela->r_offset,
1940 sym_sec->owner, sym_sec);
1941 continue;
1944 is_call = (insn[0] & 0xfd) == 0x31;
1946 if (h)
1947 val = h->root.u.def.value;
1948 else
1949 val = sym->st_value;
1950 val += irela->r_addend;
1952 if (!call_tree)
1954 struct function_info *fun;
1956 if (irela->r_addend != 0)
1958 Elf_Internal_Sym *fake = bfd_zmalloc (sizeof (*fake));
1959 if (fake == NULL)
1960 return FALSE;
1961 fake->st_value = val;
1962 fake->st_shndx
1963 = _bfd_elf_section_from_bfd_section (sym_sec->owner, sym_sec);
1964 sym = fake;
1966 if (sym)
1967 fun = maybe_insert_function (sym_sec, sym, FALSE, is_call);
1968 else
1969 fun = maybe_insert_function (sym_sec, h, TRUE, is_call);
1970 if (fun == NULL)
1971 return FALSE;
1972 if (irela->r_addend != 0
1973 && fun->u.sym != sym)
1974 free (sym);
1975 continue;
1978 caller = find_function (sec, irela->r_offset, info);
1979 if (caller == NULL)
1980 return FALSE;
1981 callee = bfd_malloc (sizeof *callee);
1982 if (callee == NULL)
1983 return FALSE;
1985 callee->fun = find_function (sym_sec, val, info);
1986 if (callee->fun == NULL)
1987 return FALSE;
1988 callee->is_tail = !is_call;
1989 if (!insert_callee (caller, callee))
1990 free (callee);
1991 else if (!is_call
1992 && !callee->fun->is_func
1993 && callee->fun->stack == 0)
1995 /* This is either a tail call or a branch from one part of
1996 the function to another, ie. hot/cold section. If the
1997 destination has been called by some other function then
1998 it is a separate function. We also assume that functions
1999 are not split across input files. */
2000 if (callee->fun->start != NULL
2001 || sec->owner != sym_sec->owner)
2003 callee->fun->start = NULL;
2004 callee->fun->is_func = TRUE;
2006 else
2007 callee->fun->start = caller;
2011 return TRUE;
2014 /* Handle something like .init or .fini, which has a piece of a function.
2015 These sections are pasted together to form a single function. */
2017 static bfd_boolean
2018 pasted_function (asection *sec, struct bfd_link_info *info)
2020 struct bfd_link_order *l;
2021 struct _spu_elf_section_data *sec_data;
2022 struct spu_elf_stack_info *sinfo;
2023 Elf_Internal_Sym *fake;
2024 struct function_info *fun, *fun_start;
2026 fake = bfd_zmalloc (sizeof (*fake));
2027 if (fake == NULL)
2028 return FALSE;
2029 fake->st_value = 0;
2030 fake->st_size = sec->size;
2031 fake->st_shndx
2032 = _bfd_elf_section_from_bfd_section (sec->owner, sec);
2033 fun = maybe_insert_function (sec, fake, FALSE, FALSE);
2034 if (!fun)
2035 return FALSE;
2037 /* Find a function immediately preceding this section. */
2038 fun_start = NULL;
2039 for (l = sec->output_section->map_head.link_order; l != NULL; l = l->next)
2041 if (l->u.indirect.section == sec)
2043 if (fun_start != NULL)
2045 if (fun_start->start)
2046 fun_start = fun_start->start;
2047 fun->start = fun_start;
2049 return TRUE;
2051 if (l->type == bfd_indirect_link_order
2052 && (sec_data = spu_elf_section_data (l->u.indirect.section)) != NULL
2053 && (sinfo = sec_data->u.i.stack_info) != NULL
2054 && sinfo->num_fun != 0)
2055 fun_start = &sinfo->fun[sinfo->num_fun - 1];
2058 info->callbacks->einfo (_("%A link_order not found\n"), sec);
2059 return FALSE;
2062 /* We're only interested in code sections. Testing SEC_IN_MEMORY excludes
2063 overlay stub sections. */
2065 static bfd_boolean
2066 interesting_section (asection *s, bfd *obfd)
2068 return (s->output_section != NULL
2069 && s->output_section->owner == obfd
2070 && ((s->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_IN_MEMORY))
2071 == (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2072 && s->size != 0);
2075 /* Map address ranges in code sections to functions. */
2077 static bfd_boolean
2078 discover_functions (bfd *output_bfd, struct bfd_link_info *info)
2080 bfd *ibfd;
2081 int bfd_idx;
2082 Elf_Internal_Sym ***psym_arr;
2083 asection ***sec_arr;
2084 bfd_boolean gaps = FALSE;
2086 bfd_idx = 0;
2087 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2088 bfd_idx++;
2090 psym_arr = bfd_zmalloc (bfd_idx * sizeof (*psym_arr));
2091 if (psym_arr == NULL)
2092 return FALSE;
2093 sec_arr = bfd_zmalloc (bfd_idx * sizeof (*sec_arr));
2094 if (sec_arr == NULL)
2095 return FALSE;
2098 for (ibfd = info->input_bfds, bfd_idx = 0;
2099 ibfd != NULL;
2100 ibfd = ibfd->link_next, bfd_idx++)
2102 extern const bfd_target bfd_elf32_spu_vec;
2103 Elf_Internal_Shdr *symtab_hdr;
2104 asection *sec;
2105 size_t symcount;
2106 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
2107 asection **psecs, **p;
2109 if (ibfd->xvec != &bfd_elf32_spu_vec)
2110 continue;
2112 /* Read all the symbols. */
2113 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2114 symcount = symtab_hdr->sh_size / symtab_hdr->sh_entsize;
2115 if (symcount == 0)
2116 continue;
2118 syms = (Elf_Internal_Sym *) symtab_hdr->contents;
2119 if (syms == NULL)
2121 syms = bfd_elf_get_elf_syms (ibfd, symtab_hdr, symcount, 0,
2122 NULL, NULL, NULL);
2123 symtab_hdr->contents = (void *) syms;
2124 if (syms == NULL)
2125 return FALSE;
2128 /* Select defined function symbols that are going to be output. */
2129 psyms = bfd_malloc ((symcount + 1) * sizeof (*psyms));
2130 if (psyms == NULL)
2131 return FALSE;
2132 psym_arr[bfd_idx] = psyms;
2133 psecs = bfd_malloc (symcount * sizeof (*psecs));
2134 if (psecs == NULL)
2135 return FALSE;
2136 sec_arr[bfd_idx] = psecs;
2137 for (psy = psyms, p = psecs, sy = syms; sy < syms + symcount; ++p, ++sy)
2138 if (ELF_ST_TYPE (sy->st_info) == STT_NOTYPE
2139 || ELF_ST_TYPE (sy->st_info) == STT_FUNC)
2141 asection *s;
2143 *p = s = bfd_section_from_elf_index (ibfd, sy->st_shndx);
2144 if (s != NULL && interesting_section (s, output_bfd))
2145 *psy++ = sy;
2147 symcount = psy - psyms;
2148 *psy = NULL;
2150 /* Sort them by section and offset within section. */
2151 sort_syms_syms = syms;
2152 sort_syms_psecs = psecs;
2153 qsort (psyms, symcount, sizeof (*psyms), sort_syms);
2155 /* Now inspect the function symbols. */
2156 for (psy = psyms; psy < psyms + symcount; )
2158 asection *s = psecs[*psy - syms];
2159 Elf_Internal_Sym **psy2;
2161 for (psy2 = psy; ++psy2 < psyms + symcount; )
2162 if (psecs[*psy2 - syms] != s)
2163 break;
2165 if (!alloc_stack_info (s, psy2 - psy))
2166 return FALSE;
2167 psy = psy2;
2170 /* First install info about properly typed and sized functions.
2171 In an ideal world this will cover all code sections, except
2172 when partitioning functions into hot and cold sections,
2173 and the horrible pasted together .init and .fini functions. */
2174 for (psy = psyms; psy < psyms + symcount; ++psy)
2176 sy = *psy;
2177 if (ELF_ST_TYPE (sy->st_info) == STT_FUNC)
2179 asection *s = psecs[sy - syms];
2180 if (!maybe_insert_function (s, sy, FALSE, TRUE))
2181 return FALSE;
2185 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2186 if (interesting_section (sec, output_bfd))
2187 gaps |= check_function_ranges (sec, info);
2190 if (gaps)
2192 /* See if we can discover more function symbols by looking at
2193 relocations. */
2194 for (ibfd = info->input_bfds, bfd_idx = 0;
2195 ibfd != NULL;
2196 ibfd = ibfd->link_next, bfd_idx++)
2198 asection *sec;
2200 if (psym_arr[bfd_idx] == NULL)
2201 continue;
2203 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2204 if (interesting_section (sec, output_bfd)
2205 && sec->reloc_count != 0)
2207 if (!mark_functions_via_relocs (sec, info, FALSE))
2208 return FALSE;
2212 for (ibfd = info->input_bfds, bfd_idx = 0;
2213 ibfd != NULL;
2214 ibfd = ibfd->link_next, bfd_idx++)
2216 Elf_Internal_Shdr *symtab_hdr;
2217 asection *sec;
2218 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
2219 asection **psecs;
2221 if ((psyms = psym_arr[bfd_idx]) == NULL)
2222 continue;
2224 psecs = sec_arr[bfd_idx];
2226 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2227 syms = (Elf_Internal_Sym *) symtab_hdr->contents;
2229 gaps = FALSE;
2230 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2231 if (interesting_section (sec, output_bfd))
2232 gaps |= check_function_ranges (sec, info);
2233 if (!gaps)
2234 continue;
2236 /* Finally, install all globals. */
2237 for (psy = psyms; (sy = *psy) != NULL; ++psy)
2239 asection *s;
2241 s = psecs[sy - syms];
2243 /* Global syms might be improperly typed functions. */
2244 if (ELF_ST_TYPE (sy->st_info) != STT_FUNC
2245 && ELF_ST_BIND (sy->st_info) == STB_GLOBAL)
2247 if (!maybe_insert_function (s, sy, FALSE, FALSE))
2248 return FALSE;
2252 /* Some of the symbols we've installed as marking the
2253 beginning of functions may have a size of zero. Extend
2254 the range of such functions to the beginning of the
2255 next symbol of interest. */
2256 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2257 if (interesting_section (sec, output_bfd))
2259 struct _spu_elf_section_data *sec_data;
2260 struct spu_elf_stack_info *sinfo;
2262 sec_data = spu_elf_section_data (sec);
2263 sinfo = sec_data->u.i.stack_info;
2264 if (sinfo != NULL)
2266 int fun_idx;
2267 bfd_vma hi = sec->size;
2269 for (fun_idx = sinfo->num_fun; --fun_idx >= 0; )
2271 sinfo->fun[fun_idx].hi = hi;
2272 hi = sinfo->fun[fun_idx].lo;
2275 /* No symbols in this section. Must be .init or .fini
2276 or something similar. */
2277 else if (!pasted_function (sec, info))
2278 return FALSE;
2283 for (ibfd = info->input_bfds, bfd_idx = 0;
2284 ibfd != NULL;
2285 ibfd = ibfd->link_next, bfd_idx++)
2287 if (psym_arr[bfd_idx] == NULL)
2288 continue;
2290 free (psym_arr[bfd_idx]);
2291 free (sec_arr[bfd_idx]);
2294 free (psym_arr);
2295 free (sec_arr);
2297 return TRUE;
2300 /* Mark nodes in the call graph that are called by some other node. */
2302 static void
2303 mark_non_root (struct function_info *fun)
2305 struct call_info *call;
2307 fun->visit1 = TRUE;
2308 for (call = fun->call_list; call; call = call->next)
2310 call->fun->non_root = TRUE;
2311 if (!call->fun->visit1)
2312 mark_non_root (call->fun);
2316 /* Remove cycles from the call graph. */
2318 static void
2319 call_graph_traverse (struct function_info *fun, struct bfd_link_info *info)
2321 struct call_info **callp, *call;
2323 fun->visit2 = TRUE;
2324 fun->marking = TRUE;
2326 callp = &fun->call_list;
2327 while ((call = *callp) != NULL)
2329 if (!call->fun->visit2)
2330 call_graph_traverse (call->fun, info);
2331 else if (call->fun->marking)
2333 const char *f1 = func_name (fun);
2334 const char *f2 = func_name (call->fun);
2336 info->callbacks->info (_("Stack analysis will ignore the call "
2337 "from %s to %s\n"),
2338 f1, f2);
2339 *callp = call->next;
2340 continue;
2342 callp = &call->next;
2344 fun->marking = FALSE;
2347 /* Populate call_list for each function. */
2349 static bfd_boolean
2350 build_call_tree (bfd *output_bfd, struct bfd_link_info *info)
2352 bfd *ibfd;
2354 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2356 extern const bfd_target bfd_elf32_spu_vec;
2357 asection *sec;
2359 if (ibfd->xvec != &bfd_elf32_spu_vec)
2360 continue;
2362 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2364 if (!interesting_section (sec, output_bfd)
2365 || sec->reloc_count == 0)
2366 continue;
2368 if (!mark_functions_via_relocs (sec, info, TRUE))
2369 return FALSE;
2372 /* Transfer call info from hot/cold section part of function
2373 to main entry. */
2374 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2376 struct _spu_elf_section_data *sec_data;
2377 struct spu_elf_stack_info *sinfo;
2379 if ((sec_data = spu_elf_section_data (sec)) != NULL
2380 && (sinfo = sec_data->u.i.stack_info) != NULL)
2382 int i;
2383 for (i = 0; i < sinfo->num_fun; ++i)
2385 if (sinfo->fun[i].start != NULL)
2387 struct call_info *call = sinfo->fun[i].call_list;
2389 while (call != NULL)
2391 struct call_info *call_next = call->next;
2392 if (!insert_callee (sinfo->fun[i].start, call))
2393 free (call);
2394 call = call_next;
2396 sinfo->fun[i].call_list = NULL;
2397 sinfo->fun[i].non_root = TRUE;
2404 /* Find the call graph root(s). */
2405 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2407 extern const bfd_target bfd_elf32_spu_vec;
2408 asection *sec;
2410 if (ibfd->xvec != &bfd_elf32_spu_vec)
2411 continue;
2413 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2415 struct _spu_elf_section_data *sec_data;
2416 struct spu_elf_stack_info *sinfo;
2418 if ((sec_data = spu_elf_section_data (sec)) != NULL
2419 && (sinfo = sec_data->u.i.stack_info) != NULL)
2421 int i;
2422 for (i = 0; i < sinfo->num_fun; ++i)
2423 if (!sinfo->fun[i].visit1)
2424 mark_non_root (&sinfo->fun[i]);
2429 /* Remove cycles from the call graph. We start from the root node(s)
2430 so that we break cycles in a reasonable place. */
2431 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2433 extern const bfd_target bfd_elf32_spu_vec;
2434 asection *sec;
2436 if (ibfd->xvec != &bfd_elf32_spu_vec)
2437 continue;
2439 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2441 struct _spu_elf_section_data *sec_data;
2442 struct spu_elf_stack_info *sinfo;
2444 if ((sec_data = spu_elf_section_data (sec)) != NULL
2445 && (sinfo = sec_data->u.i.stack_info) != NULL)
2447 int i;
2448 for (i = 0; i < sinfo->num_fun; ++i)
2449 if (!sinfo->fun[i].non_root)
2450 call_graph_traverse (&sinfo->fun[i], info);
2455 return TRUE;
2458 /* Descend the call graph for FUN, accumulating total stack required. */
2460 static bfd_vma
2461 sum_stack (struct function_info *fun,
2462 struct bfd_link_info *info,
2463 int emit_stack_syms)
2465 struct call_info *call;
2466 struct function_info *max = NULL;
2467 bfd_vma max_stack = fun->stack;
2468 bfd_vma stack;
2469 const char *f1;
2471 if (fun->visit3)
2472 return max_stack;
2474 for (call = fun->call_list; call; call = call->next)
2476 stack = sum_stack (call->fun, info, emit_stack_syms);
2477 /* Include caller stack for normal calls, don't do so for
2478 tail calls. fun->stack here is local stack usage for
2479 this function. */
2480 if (!call->is_tail)
2481 stack += fun->stack;
2482 if (max_stack < stack)
2484 max_stack = stack;
2485 max = call->fun;
2489 f1 = func_name (fun);
2490 info->callbacks->minfo (_("%s: 0x%v 0x%v\n"),
2491 f1, (bfd_vma) fun->stack, max_stack);
2493 if (fun->call_list)
2495 info->callbacks->minfo (_(" calls:\n"));
2496 for (call = fun->call_list; call; call = call->next)
2498 const char *f2 = func_name (call->fun);
2499 const char *ann1 = call->fun == max ? "*" : " ";
2500 const char *ann2 = call->is_tail ? "t" : " ";
2502 info->callbacks->minfo (_(" %s%s %s\n"), ann1, ann2, f2);
2506 /* Now fun->stack holds cumulative stack. */
2507 fun->stack = max_stack;
2508 fun->visit3 = TRUE;
2510 if (emit_stack_syms)
2512 struct spu_link_hash_table *htab = spu_hash_table (info);
2513 char *name = bfd_malloc (18 + strlen (f1));
2514 struct elf_link_hash_entry *h;
2516 if (name != NULL)
2518 if (fun->global || ELF_ST_BIND (fun->u.sym->st_info) == STB_GLOBAL)
2519 sprintf (name, "__stack_%s", f1);
2520 else
2521 sprintf (name, "__stack_%x_%s", fun->sec->id & 0xffffffff, f1);
2523 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
2524 free (name);
2525 if (h != NULL
2526 && (h->root.type == bfd_link_hash_new
2527 || h->root.type == bfd_link_hash_undefined
2528 || h->root.type == bfd_link_hash_undefweak))
2530 h->root.type = bfd_link_hash_defined;
2531 h->root.u.def.section = bfd_abs_section_ptr;
2532 h->root.u.def.value = max_stack;
2533 h->size = 0;
2534 h->type = 0;
2535 h->ref_regular = 1;
2536 h->def_regular = 1;
2537 h->ref_regular_nonweak = 1;
2538 h->forced_local = 1;
2539 h->non_elf = 0;
2544 return max_stack;
2547 /* Provide an estimate of total stack required. */
2549 static bfd_boolean
2550 spu_elf_stack_analysis (bfd *output_bfd,
2551 struct bfd_link_info *info,
2552 int emit_stack_syms)
2554 bfd *ibfd;
2555 bfd_vma max_stack = 0;
2557 if (!discover_functions (output_bfd, info))
2558 return FALSE;
2560 if (!build_call_tree (output_bfd, info))
2561 return FALSE;
2563 info->callbacks->info (_("Stack size for call graph root nodes.\n"));
2564 info->callbacks->minfo (_("\nStack size for functions. "
2565 "Annotations: '*' max stack, 't' tail call\n"));
2566 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2568 extern const bfd_target bfd_elf32_spu_vec;
2569 asection *sec;
2571 if (ibfd->xvec != &bfd_elf32_spu_vec)
2572 continue;
2574 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2576 struct _spu_elf_section_data *sec_data;
2577 struct spu_elf_stack_info *sinfo;
2579 if ((sec_data = spu_elf_section_data (sec)) != NULL
2580 && (sinfo = sec_data->u.i.stack_info) != NULL)
2582 int i;
2583 for (i = 0; i < sinfo->num_fun; ++i)
2585 if (!sinfo->fun[i].non_root)
2587 bfd_vma stack;
2588 const char *f1;
2590 stack = sum_stack (&sinfo->fun[i], info,
2591 emit_stack_syms);
2592 f1 = func_name (&sinfo->fun[i]);
2593 info->callbacks->info (_(" %s: 0x%v\n"),
2594 f1, stack);
2595 if (max_stack < stack)
2596 max_stack = stack;
2603 info->callbacks->info (_("Maximum stack required is 0x%v\n"), max_stack);
2604 return TRUE;
2607 /* Perform a final link. */
2609 static bfd_boolean
2610 spu_elf_final_link (bfd *output_bfd, struct bfd_link_info *info)
2612 struct spu_link_hash_table *htab = spu_hash_table (info);
2614 if (htab->stack_analysis
2615 && !spu_elf_stack_analysis (output_bfd, info, htab->emit_stack_syms))
2616 info->callbacks->einfo ("%X%P: stack analysis error: %E\n");
2618 return bfd_elf_final_link (output_bfd, info);
2621 /* Called when not normally emitting relocs, ie. !info->relocatable
2622 and !info->emitrelocations. Returns a count of special relocs
2623 that need to be emitted. */
2625 static unsigned int
2626 spu_elf_count_relocs (asection *sec, Elf_Internal_Rela *relocs)
2628 unsigned int count = 0;
2629 Elf_Internal_Rela *relend = relocs + sec->reloc_count;
2631 for (; relocs < relend; relocs++)
2633 int r_type = ELF32_R_TYPE (relocs->r_info);
2634 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
2635 ++count;
2638 return count;
2641 /* Apply RELOCS to CONTENTS of INPUT_SECTION from INPUT_BFD. */
2643 static bfd_boolean
2644 spu_elf_relocate_section (bfd *output_bfd,
2645 struct bfd_link_info *info,
2646 bfd *input_bfd,
2647 asection *input_section,
2648 bfd_byte *contents,
2649 Elf_Internal_Rela *relocs,
2650 Elf_Internal_Sym *local_syms,
2651 asection **local_sections)
2653 Elf_Internal_Shdr *symtab_hdr;
2654 struct elf_link_hash_entry **sym_hashes;
2655 Elf_Internal_Rela *rel, *relend;
2656 struct spu_link_hash_table *htab;
2657 bfd_boolean ret = TRUE;
2658 bfd_boolean emit_these_relocs = FALSE;
2660 htab = spu_hash_table (info);
2661 symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
2662 sym_hashes = (struct elf_link_hash_entry **) (elf_sym_hashes (input_bfd));
2664 rel = relocs;
2665 relend = relocs + input_section->reloc_count;
2666 for (; rel < relend; rel++)
2668 int r_type;
2669 reloc_howto_type *howto;
2670 unsigned long r_symndx;
2671 Elf_Internal_Sym *sym;
2672 asection *sec;
2673 struct elf_link_hash_entry *h;
2674 const char *sym_name;
2675 bfd_vma relocation;
2676 bfd_vma addend;
2677 bfd_reloc_status_type r;
2678 bfd_boolean unresolved_reloc;
2679 bfd_boolean warned;
2680 bfd_boolean branch;
2682 r_symndx = ELF32_R_SYM (rel->r_info);
2683 r_type = ELF32_R_TYPE (rel->r_info);
2684 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
2686 emit_these_relocs = TRUE;
2687 continue;
2690 howto = elf_howto_table + r_type;
2691 unresolved_reloc = FALSE;
2692 warned = FALSE;
2693 h = NULL;
2694 sym = NULL;
2695 sec = NULL;
2696 if (r_symndx < symtab_hdr->sh_info)
2698 sym = local_syms + r_symndx;
2699 sec = local_sections[r_symndx];
2700 sym_name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym, sec);
2701 relocation = _bfd_elf_rela_local_sym (output_bfd, sym, &sec, rel);
2703 else
2705 RELOC_FOR_GLOBAL_SYMBOL (info, input_bfd, input_section, rel,
2706 r_symndx, symtab_hdr, sym_hashes,
2707 h, sec, relocation,
2708 unresolved_reloc, warned);
2709 sym_name = h->root.root.string;
2712 if (sec != NULL && elf_discarded_section (sec))
2714 /* For relocs against symbols from removed linkonce sections,
2715 or sections discarded by a linker script, we just want the
2716 section contents zeroed. Avoid any special processing. */
2717 _bfd_clear_contents (howto, input_bfd, contents + rel->r_offset);
2718 rel->r_info = 0;
2719 rel->r_addend = 0;
2720 continue;
2723 if (info->relocatable)
2724 continue;
2726 if (unresolved_reloc)
2728 (*_bfd_error_handler)
2729 (_("%B(%s+0x%lx): unresolvable %s relocation against symbol `%s'"),
2730 input_bfd,
2731 bfd_get_section_name (input_bfd, input_section),
2732 (long) rel->r_offset,
2733 howto->name,
2734 sym_name);
2735 ret = FALSE;
2738 /* If this symbol is in an overlay area, we may need to relocate
2739 to the overlay stub. */
2740 addend = rel->r_addend;
2741 branch = (is_branch (contents + rel->r_offset)
2742 || is_hint (contents + rel->r_offset));
2743 if (htab->stub_sec != NULL
2744 && needs_ovl_stub (sym_name, sec, input_section, htab, branch)
2745 && (h == NULL
2746 || (h != htab->ovly_load && h != htab->ovly_return)))
2748 unsigned int ovl = 0;
2749 struct got_entry *g, **head;
2751 if (branch)
2752 ovl = (spu_elf_section_data (input_section->output_section)
2753 ->u.o.ovl_index);
2755 if (h != NULL)
2756 head = &h->got.glist;
2757 else
2758 head = elf_local_got_ents (input_bfd) + r_symndx;
2760 for (g = *head; g != NULL; g = g->next)
2761 if (g->ovl == ovl || g->ovl == 0)
2762 break;
2763 if (g == NULL)
2764 abort ();
2766 relocation = g->stub_addr;
2767 addend = 0;
2770 r = _bfd_final_link_relocate (howto,
2771 input_bfd,
2772 input_section,
2773 contents,
2774 rel->r_offset, relocation, addend);
2776 if (r != bfd_reloc_ok)
2778 const char *msg = (const char *) 0;
2780 switch (r)
2782 case bfd_reloc_overflow:
2783 if (!((*info->callbacks->reloc_overflow)
2784 (info, (h ? &h->root : NULL), sym_name, howto->name,
2785 (bfd_vma) 0, input_bfd, input_section, rel->r_offset)))
2786 return FALSE;
2787 break;
2789 case bfd_reloc_undefined:
2790 if (!((*info->callbacks->undefined_symbol)
2791 (info, sym_name, input_bfd, input_section,
2792 rel->r_offset, TRUE)))
2793 return FALSE;
2794 break;
2796 case bfd_reloc_outofrange:
2797 msg = _("internal error: out of range error");
2798 goto common_error;
2800 case bfd_reloc_notsupported:
2801 msg = _("internal error: unsupported relocation error");
2802 goto common_error;
2804 case bfd_reloc_dangerous:
2805 msg = _("internal error: dangerous error");
2806 goto common_error;
2808 default:
2809 msg = _("internal error: unknown error");
2810 /* fall through */
2812 common_error:
2813 if (!((*info->callbacks->warning)
2814 (info, msg, sym_name, input_bfd, input_section,
2815 rel->r_offset)))
2816 return FALSE;
2817 break;
2822 if (ret
2823 && emit_these_relocs
2824 && !info->relocatable
2825 && !info->emitrelocations)
2827 Elf_Internal_Rela *wrel;
2828 Elf_Internal_Shdr *rel_hdr;
2830 wrel = rel = relocs;
2831 relend = relocs + input_section->reloc_count;
2832 for (; rel < relend; rel++)
2834 int r_type;
2836 r_type = ELF32_R_TYPE (rel->r_info);
2837 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
2838 *wrel++ = *rel;
2840 input_section->reloc_count = wrel - relocs;
2841 /* Backflips for _bfd_elf_link_output_relocs. */
2842 rel_hdr = &elf_section_data (input_section)->rel_hdr;
2843 rel_hdr->sh_size = input_section->reloc_count * rel_hdr->sh_entsize;
2844 ret = 2;
2847 return ret;
2850 /* Adjust _SPUEAR_ syms to point at their overlay stubs. */
2852 static bfd_boolean
2853 spu_elf_output_symbol_hook (struct bfd_link_info *info,
2854 const char *sym_name ATTRIBUTE_UNUSED,
2855 Elf_Internal_Sym *sym,
2856 asection *sym_sec ATTRIBUTE_UNUSED,
2857 struct elf_link_hash_entry *h)
2859 struct spu_link_hash_table *htab = spu_hash_table (info);
2861 if (!info->relocatable
2862 && htab->stub_sec != NULL
2863 && h != NULL
2864 && (h->root.type == bfd_link_hash_defined
2865 || h->root.type == bfd_link_hash_defweak)
2866 && h->def_regular
2867 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0)
2869 struct got_entry *g = h->got.glist;
2871 if (g != NULL && g->ovl == 0)
2873 sym->st_shndx = (_bfd_elf_section_from_bfd_section
2874 (htab->stub_sec[0]->output_section->owner,
2875 htab->stub_sec[0]->output_section));
2876 sym->st_value = g->stub_addr;
2880 return TRUE;
2883 static int spu_plugin = 0;
2885 void
2886 spu_elf_plugin (int val)
2888 spu_plugin = val;
2891 /* Set ELF header e_type for plugins. */
2893 static void
2894 spu_elf_post_process_headers (bfd *abfd,
2895 struct bfd_link_info *info ATTRIBUTE_UNUSED)
2897 if (spu_plugin)
2899 Elf_Internal_Ehdr *i_ehdrp = elf_elfheader (abfd);
2901 i_ehdrp->e_type = ET_DYN;
2905 /* We may add an extra PT_LOAD segment for .toe. We also need extra
2906 segments for overlays. */
2908 static int
2909 spu_elf_additional_program_headers (bfd *abfd, struct bfd_link_info *info)
2911 struct spu_link_hash_table *htab = spu_hash_table (info);
2912 int extra = htab->num_overlays;
2913 asection *sec;
2915 if (extra)
2916 ++extra;
2918 sec = bfd_get_section_by_name (abfd, ".toe");
2919 if (sec != NULL && (sec->flags & SEC_LOAD) != 0)
2920 ++extra;
2922 return extra;
2925 /* Remove .toe section from other PT_LOAD segments and put it in
2926 a segment of its own. Put overlays in separate segments too. */
2928 static bfd_boolean
2929 spu_elf_modify_segment_map (bfd *abfd, struct bfd_link_info *info)
2931 asection *toe, *s;
2932 struct elf_segment_map *m;
2933 unsigned int i;
2935 if (info == NULL)
2936 return TRUE;
2938 toe = bfd_get_section_by_name (abfd, ".toe");
2939 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
2940 if (m->p_type == PT_LOAD && m->count > 1)
2941 for (i = 0; i < m->count; i++)
2942 if ((s = m->sections[i]) == toe
2943 || spu_elf_section_data (s)->u.o.ovl_index != 0)
2945 struct elf_segment_map *m2;
2946 bfd_vma amt;
2948 if (i + 1 < m->count)
2950 amt = sizeof (struct elf_segment_map);
2951 amt += (m->count - (i + 2)) * sizeof (m->sections[0]);
2952 m2 = bfd_zalloc (abfd, amt);
2953 if (m2 == NULL)
2954 return FALSE;
2955 m2->count = m->count - (i + 1);
2956 memcpy (m2->sections, m->sections + i + 1,
2957 m2->count * sizeof (m->sections[0]));
2958 m2->p_type = PT_LOAD;
2959 m2->next = m->next;
2960 m->next = m2;
2962 m->count = 1;
2963 if (i != 0)
2965 m->count = i;
2966 amt = sizeof (struct elf_segment_map);
2967 m2 = bfd_zalloc (abfd, amt);
2968 if (m2 == NULL)
2969 return FALSE;
2970 m2->p_type = PT_LOAD;
2971 m2->count = 1;
2972 m2->sections[0] = s;
2973 m2->next = m->next;
2974 m->next = m2;
2976 break;
2979 return TRUE;
2982 /* Check that all loadable section VMAs lie in the range
2983 LO .. HI inclusive. */
2985 asection *
2986 spu_elf_check_vma (bfd *abfd, bfd_vma lo, bfd_vma hi)
2988 struct elf_segment_map *m;
2989 unsigned int i;
2991 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
2992 if (m->p_type == PT_LOAD)
2993 for (i = 0; i < m->count; i++)
2994 if (m->sections[i]->size != 0
2995 && (m->sections[i]->vma < lo
2996 || m->sections[i]->vma > hi
2997 || m->sections[i]->vma + m->sections[i]->size - 1 > hi))
2998 return m->sections[i];
3000 return NULL;
3003 /* Tweak the section type of .note.spu_name. */
3005 static bfd_boolean
3006 spu_elf_fake_sections (bfd *obfd ATTRIBUTE_UNUSED,
3007 Elf_Internal_Shdr *hdr,
3008 asection *sec)
3010 if (strcmp (sec->name, SPU_PTNOTE_SPUNAME) == 0)
3011 hdr->sh_type = SHT_NOTE;
3012 return TRUE;
3015 /* Tweak phdrs before writing them out. */
3017 static int
3018 spu_elf_modify_program_headers (bfd *abfd, struct bfd_link_info *info)
3020 const struct elf_backend_data *bed;
3021 struct elf_obj_tdata *tdata;
3022 Elf_Internal_Phdr *phdr, *last;
3023 struct spu_link_hash_table *htab;
3024 unsigned int count;
3025 unsigned int i;
3027 if (info == NULL)
3028 return TRUE;
3030 bed = get_elf_backend_data (abfd);
3031 tdata = elf_tdata (abfd);
3032 phdr = tdata->phdr;
3033 count = tdata->program_header_size / bed->s->sizeof_phdr;
3034 htab = spu_hash_table (info);
3035 if (htab->num_overlays != 0)
3037 struct elf_segment_map *m;
3038 unsigned int o;
3040 for (i = 0, m = elf_tdata (abfd)->segment_map; m; ++i, m = m->next)
3041 if (m->count != 0
3042 && (o = spu_elf_section_data (m->sections[0])->u.o.ovl_index) != 0)
3044 /* Mark this as an overlay header. */
3045 phdr[i].p_flags |= PF_OVERLAY;
3047 if (htab->ovtab != NULL && htab->ovtab->size != 0)
3049 bfd_byte *p = htab->ovtab->contents;
3050 unsigned int off = o * 16 + 8;
3052 /* Write file_off into _ovly_table. */
3053 bfd_put_32 (htab->ovtab->owner, phdr[i].p_offset, p + off);
3058 /* Round up p_filesz and p_memsz of PT_LOAD segments to multiples
3059 of 16. This should always be possible when using the standard
3060 linker scripts, but don't create overlapping segments if
3061 someone is playing games with linker scripts. */
3062 last = NULL;
3063 for (i = count; i-- != 0; )
3064 if (phdr[i].p_type == PT_LOAD)
3066 unsigned adjust;
3068 adjust = -phdr[i].p_filesz & 15;
3069 if (adjust != 0
3070 && last != NULL
3071 && phdr[i].p_offset + phdr[i].p_filesz > last->p_offset - adjust)
3072 break;
3074 adjust = -phdr[i].p_memsz & 15;
3075 if (adjust != 0
3076 && last != NULL
3077 && phdr[i].p_filesz != 0
3078 && phdr[i].p_vaddr + phdr[i].p_memsz > last->p_vaddr - adjust
3079 && phdr[i].p_vaddr + phdr[i].p_memsz <= last->p_vaddr)
3080 break;
3082 if (phdr[i].p_filesz != 0)
3083 last = &phdr[i];
3086 if (i == (unsigned int) -1)
3087 for (i = count; i-- != 0; )
3088 if (phdr[i].p_type == PT_LOAD)
3090 unsigned adjust;
3092 adjust = -phdr[i].p_filesz & 15;
3093 phdr[i].p_filesz += adjust;
3095 adjust = -phdr[i].p_memsz & 15;
3096 phdr[i].p_memsz += adjust;
3099 return TRUE;
3102 #define TARGET_BIG_SYM bfd_elf32_spu_vec
3103 #define TARGET_BIG_NAME "elf32-spu"
3104 #define ELF_ARCH bfd_arch_spu
3105 #define ELF_MACHINE_CODE EM_SPU
3106 /* This matches the alignment need for DMA. */
3107 #define ELF_MAXPAGESIZE 0x80
3108 #define elf_backend_rela_normal 1
3109 #define elf_backend_can_gc_sections 1
3111 #define bfd_elf32_bfd_reloc_type_lookup spu_elf_reloc_type_lookup
3112 #define bfd_elf32_bfd_reloc_name_lookup spu_elf_reloc_name_lookup
3113 #define elf_info_to_howto spu_elf_info_to_howto
3114 #define elf_backend_count_relocs spu_elf_count_relocs
3115 #define elf_backend_relocate_section spu_elf_relocate_section
3116 #define elf_backend_symbol_processing spu_elf_backend_symbol_processing
3117 #define elf_backend_link_output_symbol_hook spu_elf_output_symbol_hook
3118 #define bfd_elf32_new_section_hook spu_elf_new_section_hook
3119 #define bfd_elf32_bfd_link_hash_table_create spu_elf_link_hash_table_create
3121 #define elf_backend_additional_program_headers spu_elf_additional_program_headers
3122 #define elf_backend_modify_segment_map spu_elf_modify_segment_map
3123 #define elf_backend_modify_program_headers spu_elf_modify_program_headers
3124 #define elf_backend_post_process_headers spu_elf_post_process_headers
3125 #define elf_backend_fake_sections spu_elf_fake_sections
3126 #define elf_backend_special_sections spu_elf_special_sections
3127 #define bfd_elf32_bfd_final_link spu_elf_final_link
3129 #include "elf32-target.h"