Define __ASSUME_AT_RANDOM.
[glibc/history.git] / elf / dl-load.c
blob0b896d9b47ac58bc86b23629a48e0ef7b92841cb
1 /* Map in a shared object's segments from the file.
2 Copyright (C) 1995-2005, 2006, 2007, 2009 Free Software Foundation, Inc.
3 This file is part of the GNU C Library.
5 The GNU C Library is free software; you can redistribute it and/or
6 modify it under the terms of the GNU Lesser General Public
7 License as published by the Free Software Foundation; either
8 version 2.1 of the License, or (at your option) any later version.
10 The GNU C Library is distributed in the hope that it will be useful,
11 but WITHOUT ANY WARRANTY; without even the implied warranty of
12 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 Lesser General Public License for more details.
15 You should have received a copy of the GNU Lesser General Public
16 License along with the GNU C Library; if not, write to the Free
17 Software Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
18 02111-1307 USA. */
20 #include <elf.h>
21 #include <errno.h>
22 #include <fcntl.h>
23 #include <libintl.h>
24 #include <stdbool.h>
25 #include <stdlib.h>
26 #include <string.h>
27 #include <unistd.h>
28 #include <ldsodefs.h>
29 #include <bits/wordsize.h>
30 #include <sys/mman.h>
31 #include <sys/param.h>
32 #include <sys/stat.h>
33 #include <sys/types.h>
34 #include "dynamic-link.h"
35 #include <abi-tag.h>
36 #include <stackinfo.h>
37 #include <caller.h>
38 #include <sysdep.h>
40 #include <dl-dst.h>
42 /* On some systems, no flag bits are given to specify file mapping. */
43 #ifndef MAP_FILE
44 # define MAP_FILE 0
45 #endif
47 /* The right way to map in the shared library files is MAP_COPY, which
48 makes a virtual copy of the data at the time of the mmap call; this
49 guarantees the mapped pages will be consistent even if the file is
50 overwritten. Some losing VM systems like Linux's lack MAP_COPY. All we
51 get is MAP_PRIVATE, which copies each page when it is modified; this
52 means if the file is overwritten, we may at some point get some pages
53 from the new version after starting with pages from the old version.
55 To make up for the lack and avoid the overwriting problem,
56 what Linux does have is MAP_DENYWRITE. This prevents anyone
57 from modifying the file while we have it mapped. */
58 #ifndef MAP_COPY
59 # ifdef MAP_DENYWRITE
60 # define MAP_COPY (MAP_PRIVATE | MAP_DENYWRITE)
61 # else
62 # define MAP_COPY MAP_PRIVATE
63 # endif
64 #endif
66 /* Some systems link their relocatable objects for another base address
67 than 0. We want to know the base address for these such that we can
68 subtract this address from the segment addresses during mapping.
69 This results in a more efficient address space usage. Defaults to
70 zero for almost all systems. */
71 #ifndef MAP_BASE_ADDR
72 # define MAP_BASE_ADDR(l) 0
73 #endif
76 #include <endian.h>
77 #if BYTE_ORDER == BIG_ENDIAN
78 # define byteorder ELFDATA2MSB
79 #elif BYTE_ORDER == LITTLE_ENDIAN
80 # define byteorder ELFDATA2LSB
81 #else
82 # error "Unknown BYTE_ORDER " BYTE_ORDER
83 # define byteorder ELFDATANONE
84 #endif
86 #define STRING(x) __STRING (x)
88 /* Handle situations where we have a preferred location in memory for
89 the shared objects. */
90 #ifdef ELF_PREFERRED_ADDRESS_DATA
91 ELF_PREFERRED_ADDRESS_DATA;
92 #endif
93 #ifndef ELF_PREFERRED_ADDRESS
94 # define ELF_PREFERRED_ADDRESS(loader, maplength, mapstartpref) (mapstartpref)
95 #endif
96 #ifndef ELF_FIXED_ADDRESS
97 # define ELF_FIXED_ADDRESS(loader, mapstart) ((void) 0)
98 #endif
101 int __stack_prot attribute_hidden attribute_relro
102 #if _STACK_GROWS_DOWN && defined PROT_GROWSDOWN
103 = PROT_GROWSDOWN;
104 #elif _STACK_GROWS_UP && defined PROT_GROWSUP
105 = PROT_GROWSUP;
106 #else
107 = 0;
108 #endif
111 /* Type for the buffer we put the ELF header and hopefully the program
112 header. This buffer does not really have to be too large. In most
113 cases the program header follows the ELF header directly. If this
114 is not the case all bets are off and we can make the header
115 arbitrarily large and still won't get it read. This means the only
116 question is how large are the ELF and program header combined. The
117 ELF header 32-bit files is 52 bytes long and in 64-bit files is 64
118 bytes long. Each program header entry is again 32 and 56 bytes
119 long respectively. I.e., even with a file which has 10 program
120 header entries we only have to read 372B/624B respectively. Add to
121 this a bit of margin for program notes and reading 512B and 832B
122 for 32-bit and 64-bit files respecitvely is enough. If this
123 heuristic should really fail for some file the code in
124 `_dl_map_object_from_fd' knows how to recover. */
125 struct filebuf
127 ssize_t len;
128 #if __WORDSIZE == 32
129 # define FILEBUF_SIZE 512
130 #else
131 # define FILEBUF_SIZE 832
132 #endif
133 char buf[FILEBUF_SIZE] __attribute__ ((aligned (__alignof (ElfW(Ehdr)))));
136 /* This is the decomposed LD_LIBRARY_PATH search path. */
137 static struct r_search_path_struct env_path_list attribute_relro;
139 /* List of the hardware capabilities we might end up using. */
140 static const struct r_strlenpair *capstr attribute_relro;
141 static size_t ncapstr attribute_relro;
142 static size_t max_capstrlen attribute_relro;
145 /* Get the generated information about the trusted directories. */
146 #include "trusted-dirs.h"
148 static const char system_dirs[] = SYSTEM_DIRS;
149 static const size_t system_dirs_len[] =
151 SYSTEM_DIRS_LEN
153 #define nsystem_dirs_len \
154 (sizeof (system_dirs_len) / sizeof (system_dirs_len[0]))
157 /* Local version of `strdup' function. */
158 static char *
159 local_strdup (const char *s)
161 size_t len = strlen (s) + 1;
162 void *new = malloc (len);
164 if (new == NULL)
165 return NULL;
167 return (char *) memcpy (new, s, len);
171 static size_t
172 is_dst (const char *start, const char *name, const char *str,
173 int is_path, int secure)
175 size_t len;
176 bool is_curly = false;
178 if (name[0] == '{')
180 is_curly = true;
181 ++name;
184 len = 0;
185 while (name[len] == str[len] && name[len] != '\0')
186 ++len;
188 if (is_curly)
190 if (name[len] != '}')
191 return 0;
193 /* Point again at the beginning of the name. */
194 --name;
195 /* Skip over closing curly brace and adjust for the --name. */
196 len += 2;
198 else if (name[len] != '\0' && name[len] != '/'
199 && (!is_path || name[len] != ':'))
200 return 0;
202 if (__builtin_expect (secure, 0)
203 && ((name[len] != '\0' && (!is_path || name[len] != ':'))
204 || (name != start + 1 && (!is_path || name[-2] != ':'))))
205 return 0;
207 return len;
211 size_t
212 _dl_dst_count (const char *name, int is_path)
214 const char *const start = name;
215 size_t cnt = 0;
219 size_t len;
221 /* $ORIGIN is not expanded for SUID/GUID programs (except if it
222 is $ORIGIN alone) and it must always appear first in path. */
223 ++name;
224 if ((len = is_dst (start, name, "ORIGIN", is_path,
225 INTUSE(__libc_enable_secure))) != 0
226 || (len = is_dst (start, name, "PLATFORM", is_path, 0)) != 0
227 || (len = is_dst (start, name, "LIB", is_path, 0)) != 0)
228 ++cnt;
230 name = strchr (name + len, '$');
232 while (name != NULL);
234 return cnt;
238 char *
239 _dl_dst_substitute (struct link_map *l, const char *name, char *result,
240 int is_path)
242 const char *const start = name;
243 char *last_elem, *wp;
245 /* Now fill the result path. While copying over the string we keep
246 track of the start of the last path element. When we come accross
247 a DST we copy over the value or (if the value is not available)
248 leave the entire path element out. */
249 last_elem = wp = result;
253 if (__builtin_expect (*name == '$', 0))
255 const char *repl = NULL;
256 size_t len;
258 ++name;
259 if ((len = is_dst (start, name, "ORIGIN", is_path,
260 INTUSE(__libc_enable_secure))) != 0)
262 #ifndef SHARED
263 if (l == NULL)
264 repl = _dl_get_origin ();
265 else
266 #endif
267 repl = l->l_origin;
269 else if ((len = is_dst (start, name, "PLATFORM", is_path, 0)) != 0)
270 repl = GLRO(dl_platform);
271 else if ((len = is_dst (start, name, "LIB", is_path, 0)) != 0)
272 repl = DL_DST_LIB;
274 if (repl != NULL && repl != (const char *) -1)
276 wp = __stpcpy (wp, repl);
277 name += len;
279 else if (len > 1)
281 /* We cannot use this path element, the value of the
282 replacement is unknown. */
283 wp = last_elem;
284 name += len;
285 while (*name != '\0' && (!is_path || *name != ':'))
286 ++name;
288 else
289 /* No DST we recognize. */
290 *wp++ = '$';
292 else
294 *wp++ = *name++;
295 if (is_path && *name == ':')
296 last_elem = wp;
299 while (*name != '\0');
301 *wp = '\0';
303 return result;
307 /* Return copy of argument with all recognized dynamic string tokens
308 ($ORIGIN and $PLATFORM for now) replaced. On some platforms it
309 might not be possible to determine the path from which the object
310 belonging to the map is loaded. In this case the path element
311 containing $ORIGIN is left out. */
312 static char *
313 expand_dynamic_string_token (struct link_map *l, const char *s)
315 /* We make two runs over the string. First we determine how large the
316 resulting string is and then we copy it over. Since this is now
317 frequently executed operation we are looking here not for performance
318 but rather for code size. */
319 size_t cnt;
320 size_t total;
321 char *result;
323 /* Determine the number of DST elements. */
324 cnt = DL_DST_COUNT (s, 1);
326 /* If we do not have to replace anything simply copy the string. */
327 if (__builtin_expect (cnt, 0) == 0)
328 return local_strdup (s);
330 /* Determine the length of the substituted string. */
331 total = DL_DST_REQUIRED (l, s, strlen (s), cnt);
333 /* Allocate the necessary memory. */
334 result = (char *) malloc (total + 1);
335 if (result == NULL)
336 return NULL;
338 return _dl_dst_substitute (l, s, result, 1);
342 /* Add `name' to the list of names for a particular shared object.
343 `name' is expected to have been allocated with malloc and will
344 be freed if the shared object already has this name.
345 Returns false if the object already had this name. */
346 static void
347 internal_function
348 add_name_to_object (struct link_map *l, const char *name)
350 struct libname_list *lnp, *lastp;
351 struct libname_list *newname;
352 size_t name_len;
354 lastp = NULL;
355 for (lnp = l->l_libname; lnp != NULL; lastp = lnp, lnp = lnp->next)
356 if (strcmp (name, lnp->name) == 0)
357 return;
359 name_len = strlen (name) + 1;
360 newname = (struct libname_list *) malloc (sizeof *newname + name_len);
361 if (newname == NULL)
363 /* No more memory. */
364 _dl_signal_error (ENOMEM, name, NULL, N_("cannot allocate name record"));
365 return;
367 /* The object should have a libname set from _dl_new_object. */
368 assert (lastp != NULL);
370 newname->name = memcpy (newname + 1, name, name_len);
371 newname->next = NULL;
372 newname->dont_free = 0;
373 lastp->next = newname;
376 /* Standard search directories. */
377 static struct r_search_path_struct rtld_search_dirs attribute_relro;
379 static size_t max_dirnamelen;
381 static struct r_search_path_elem **
382 fillin_rpath (char *rpath, struct r_search_path_elem **result, const char *sep,
383 int check_trusted, const char *what, const char *where)
385 char *cp;
386 size_t nelems = 0;
388 while ((cp = __strsep (&rpath, sep)) != NULL)
390 struct r_search_path_elem *dirp;
391 size_t len = strlen (cp);
393 /* `strsep' can pass an empty string. This has to be
394 interpreted as `use the current directory'. */
395 if (len == 0)
397 static const char curwd[] = "./";
398 cp = (char *) curwd;
401 /* Remove trailing slashes (except for "/"). */
402 while (len > 1 && cp[len - 1] == '/')
403 --len;
405 /* Now add one if there is none so far. */
406 if (len > 0 && cp[len - 1] != '/')
407 cp[len++] = '/';
409 /* Make sure we don't use untrusted directories if we run SUID. */
410 if (__builtin_expect (check_trusted, 0))
412 const char *trun = system_dirs;
413 size_t idx;
414 int unsecure = 1;
416 /* All trusted directories must be complete names. */
417 if (cp[0] == '/')
419 for (idx = 0; idx < nsystem_dirs_len; ++idx)
421 if (len == system_dirs_len[idx]
422 && memcmp (trun, cp, len) == 0)
424 /* Found it. */
425 unsecure = 0;
426 break;
429 trun += system_dirs_len[idx] + 1;
433 if (unsecure)
434 /* Simply drop this directory. */
435 continue;
438 /* See if this directory is already known. */
439 for (dirp = GL(dl_all_dirs); dirp != NULL; dirp = dirp->next)
440 if (dirp->dirnamelen == len && memcmp (cp, dirp->dirname, len) == 0)
441 break;
443 if (dirp != NULL)
445 /* It is available, see whether it's on our own list. */
446 size_t cnt;
447 for (cnt = 0; cnt < nelems; ++cnt)
448 if (result[cnt] == dirp)
449 break;
451 if (cnt == nelems)
452 result[nelems++] = dirp;
454 else
456 size_t cnt;
457 enum r_dir_status init_val;
458 size_t where_len = where ? strlen (where) + 1 : 0;
460 /* It's a new directory. Create an entry and add it. */
461 dirp = (struct r_search_path_elem *)
462 malloc (sizeof (*dirp) + ncapstr * sizeof (enum r_dir_status)
463 + where_len + len + 1);
464 if (dirp == NULL)
465 _dl_signal_error (ENOMEM, NULL, NULL,
466 N_("cannot create cache for search path"));
468 dirp->dirname = ((char *) dirp + sizeof (*dirp)
469 + ncapstr * sizeof (enum r_dir_status));
470 *((char *) __mempcpy ((char *) dirp->dirname, cp, len)) = '\0';
471 dirp->dirnamelen = len;
473 if (len > max_dirnamelen)
474 max_dirnamelen = len;
476 /* We have to make sure all the relative directories are
477 never ignored. The current directory might change and
478 all our saved information would be void. */
479 init_val = cp[0] != '/' ? existing : unknown;
480 for (cnt = 0; cnt < ncapstr; ++cnt)
481 dirp->status[cnt] = init_val;
483 dirp->what = what;
484 if (__builtin_expect (where != NULL, 1))
485 dirp->where = memcpy ((char *) dirp + sizeof (*dirp) + len + 1
486 + (ncapstr * sizeof (enum r_dir_status)),
487 where, where_len);
488 else
489 dirp->where = NULL;
491 dirp->next = GL(dl_all_dirs);
492 GL(dl_all_dirs) = dirp;
494 /* Put it in the result array. */
495 result[nelems++] = dirp;
499 /* Terminate the array. */
500 result[nelems] = NULL;
502 return result;
506 static bool
507 internal_function
508 decompose_rpath (struct r_search_path_struct *sps,
509 const char *rpath, struct link_map *l, const char *what)
511 /* Make a copy we can work with. */
512 const char *where = l->l_name;
513 char *copy;
514 char *cp;
515 struct r_search_path_elem **result;
516 size_t nelems;
517 /* Initialize to please the compiler. */
518 const char *errstring = NULL;
520 /* First see whether we must forget the RUNPATH and RPATH from this
521 object. */
522 if (__builtin_expect (GLRO(dl_inhibit_rpath) != NULL, 0)
523 && !INTUSE(__libc_enable_secure))
525 const char *inhp = GLRO(dl_inhibit_rpath);
529 const char *wp = where;
531 while (*inhp == *wp && *wp != '\0')
533 ++inhp;
534 ++wp;
537 if (*wp == '\0' && (*inhp == '\0' || *inhp == ':'))
539 /* This object is on the list of objects for which the
540 RUNPATH and RPATH must not be used. */
541 sps->dirs = (void *) -1;
542 return false;
545 while (*inhp != '\0')
546 if (*inhp++ == ':')
547 break;
549 while (*inhp != '\0');
552 /* Make a writable copy. At the same time expand possible dynamic
553 string tokens. */
554 copy = expand_dynamic_string_token (l, rpath);
555 if (copy == NULL)
557 errstring = N_("cannot create RUNPATH/RPATH copy");
558 goto signal_error;
561 /* Count the number of necessary elements in the result array. */
562 nelems = 0;
563 for (cp = copy; *cp != '\0'; ++cp)
564 if (*cp == ':')
565 ++nelems;
567 /* Allocate room for the result. NELEMS + 1 is an upper limit for the
568 number of necessary entries. */
569 result = (struct r_search_path_elem **) malloc ((nelems + 1 + 1)
570 * sizeof (*result));
571 if (result == NULL)
573 free (copy);
574 errstring = N_("cannot create cache for search path");
575 signal_error:
576 _dl_signal_error (ENOMEM, NULL, NULL, errstring);
579 fillin_rpath (copy, result, ":", 0, what, where);
581 /* Free the copied RPATH string. `fillin_rpath' make own copies if
582 necessary. */
583 free (copy);
585 sps->dirs = result;
586 /* The caller will change this value if we haven't used a real malloc. */
587 sps->malloced = 1;
588 return true;
591 /* Make sure cached path information is stored in *SP
592 and return true if there are any paths to search there. */
593 static bool
594 cache_rpath (struct link_map *l,
595 struct r_search_path_struct *sp,
596 int tag,
597 const char *what)
599 if (sp->dirs == (void *) -1)
600 return false;
602 if (sp->dirs != NULL)
603 return true;
605 if (l->l_info[tag] == NULL)
607 /* There is no path. */
608 sp->dirs = (void *) -1;
609 return false;
612 /* Make sure the cache information is available. */
613 return decompose_rpath (sp, (const char *) (D_PTR (l, l_info[DT_STRTAB])
614 + l->l_info[tag]->d_un.d_val),
615 l, what);
619 void
620 internal_function
621 _dl_init_paths (const char *llp)
623 size_t idx;
624 const char *strp;
625 struct r_search_path_elem *pelem, **aelem;
626 size_t round_size;
627 #ifdef SHARED
628 struct link_map *l;
629 #endif
630 /* Initialize to please the compiler. */
631 const char *errstring = NULL;
633 /* Fill in the information about the application's RPATH and the
634 directories addressed by the LD_LIBRARY_PATH environment variable. */
636 /* Get the capabilities. */
637 capstr = _dl_important_hwcaps (GLRO(dl_platform), GLRO(dl_platformlen),
638 &ncapstr, &max_capstrlen);
640 /* First set up the rest of the default search directory entries. */
641 aelem = rtld_search_dirs.dirs = (struct r_search_path_elem **)
642 malloc ((nsystem_dirs_len + 1) * sizeof (struct r_search_path_elem *));
643 if (rtld_search_dirs.dirs == NULL)
645 errstring = N_("cannot create search path array");
646 signal_error:
647 _dl_signal_error (ENOMEM, NULL, NULL, errstring);
650 round_size = ((2 * sizeof (struct r_search_path_elem) - 1
651 + ncapstr * sizeof (enum r_dir_status))
652 / sizeof (struct r_search_path_elem));
654 rtld_search_dirs.dirs[0] = (struct r_search_path_elem *)
655 malloc ((sizeof (system_dirs) / sizeof (system_dirs[0]))
656 * round_size * sizeof (struct r_search_path_elem));
657 if (rtld_search_dirs.dirs[0] == NULL)
659 errstring = N_("cannot create cache for search path");
660 goto signal_error;
663 rtld_search_dirs.malloced = 0;
664 pelem = GL(dl_all_dirs) = rtld_search_dirs.dirs[0];
665 strp = system_dirs;
666 idx = 0;
670 size_t cnt;
672 *aelem++ = pelem;
674 pelem->what = "system search path";
675 pelem->where = NULL;
677 pelem->dirname = strp;
678 pelem->dirnamelen = system_dirs_len[idx];
679 strp += system_dirs_len[idx] + 1;
681 /* System paths must be absolute. */
682 assert (pelem->dirname[0] == '/');
683 for (cnt = 0; cnt < ncapstr; ++cnt)
684 pelem->status[cnt] = unknown;
686 pelem->next = (++idx == nsystem_dirs_len ? NULL : (pelem + round_size));
688 pelem += round_size;
690 while (idx < nsystem_dirs_len);
692 max_dirnamelen = SYSTEM_DIRS_MAX_LEN;
693 *aelem = NULL;
695 #ifdef SHARED
696 /* This points to the map of the main object. */
697 l = GL(dl_ns)[LM_ID_BASE]._ns_loaded;
698 if (l != NULL)
700 assert (l->l_type != lt_loaded);
702 if (l->l_info[DT_RUNPATH])
704 /* Allocate room for the search path and fill in information
705 from RUNPATH. */
706 decompose_rpath (&l->l_runpath_dirs,
707 (const void *) (D_PTR (l, l_info[DT_STRTAB])
708 + l->l_info[DT_RUNPATH]->d_un.d_val),
709 l, "RUNPATH");
711 /* The RPATH is ignored. */
712 l->l_rpath_dirs.dirs = (void *) -1;
714 else
716 l->l_runpath_dirs.dirs = (void *) -1;
718 if (l->l_info[DT_RPATH])
720 /* Allocate room for the search path and fill in information
721 from RPATH. */
722 decompose_rpath (&l->l_rpath_dirs,
723 (const void *) (D_PTR (l, l_info[DT_STRTAB])
724 + l->l_info[DT_RPATH]->d_un.d_val),
725 l, "RPATH");
726 l->l_rpath_dirs.malloced = 0;
728 else
729 l->l_rpath_dirs.dirs = (void *) -1;
732 #endif /* SHARED */
734 if (llp != NULL && *llp != '\0')
736 size_t nllp;
737 const char *cp = llp;
738 char *llp_tmp;
740 #ifdef SHARED
741 /* Expand DSTs. */
742 size_t cnt = DL_DST_COUNT (llp, 1);
743 if (__builtin_expect (cnt == 0, 1))
744 llp_tmp = strdupa (llp);
745 else
747 /* Determine the length of the substituted string. */
748 size_t total = DL_DST_REQUIRED (l, llp, strlen (llp), cnt);
750 /* Allocate the necessary memory. */
751 llp_tmp = (char *) alloca (total + 1);
752 llp_tmp = _dl_dst_substitute (l, llp, llp_tmp, 1);
754 #else
755 llp_tmp = strdupa (llp);
756 #endif
758 /* Decompose the LD_LIBRARY_PATH contents. First determine how many
759 elements it has. */
760 nllp = 1;
761 while (*cp)
763 if (*cp == ':' || *cp == ';')
764 ++nllp;
765 ++cp;
768 env_path_list.dirs = (struct r_search_path_elem **)
769 malloc ((nllp + 1) * sizeof (struct r_search_path_elem *));
770 if (env_path_list.dirs == NULL)
772 errstring = N_("cannot create cache for search path");
773 goto signal_error;
776 (void) fillin_rpath (llp_tmp, env_path_list.dirs, ":;",
777 INTUSE(__libc_enable_secure), "LD_LIBRARY_PATH",
778 NULL);
780 if (env_path_list.dirs[0] == NULL)
782 free (env_path_list.dirs);
783 env_path_list.dirs = (void *) -1;
786 env_path_list.malloced = 0;
788 else
789 env_path_list.dirs = (void *) -1;
791 /* Remember the last search directory added at startup. */
792 GLRO(dl_init_all_dirs) = GL(dl_all_dirs);
796 static void
797 __attribute__ ((noreturn, noinline))
798 lose (int code, int fd, const char *name, char *realname, struct link_map *l,
799 const char *msg, struct r_debug *r)
801 /* The file might already be closed. */
802 if (fd != -1)
803 (void) __close (fd);
804 if (l != NULL)
806 /* Remove the stillborn object from the list and free it. */
807 assert (l->l_next == NULL);
808 if (l->l_prev == NULL)
809 /* No other module loaded. This happens only in the static library,
810 or in rtld under --verify. */
811 GL(dl_ns)[l->l_ns]._ns_loaded = NULL;
812 else
813 l->l_prev->l_next = NULL;
814 --GL(dl_ns)[l->l_ns]._ns_nloaded;
815 free (l);
817 free (realname);
819 if (r != NULL)
821 r->r_state = RT_CONSISTENT;
822 _dl_debug_state ();
825 _dl_signal_error (code, name, NULL, msg);
829 /* Map in the shared object NAME, actually located in REALNAME, and already
830 opened on FD. */
832 #ifndef EXTERNAL_MAP_FROM_FD
833 static
834 #endif
835 struct link_map *
836 _dl_map_object_from_fd (const char *name, int fd, struct filebuf *fbp,
837 char *realname, struct link_map *loader, int l_type,
838 int mode, void **stack_endp, Lmid_t nsid)
840 struct link_map *l = NULL;
841 const ElfW(Ehdr) *header;
842 const ElfW(Phdr) *phdr;
843 const ElfW(Phdr) *ph;
844 size_t maplength;
845 int type;
846 struct stat64 st;
847 /* Initialize to keep the compiler happy. */
848 const char *errstring = NULL;
849 int errval = 0;
850 struct r_debug *r = _dl_debug_initialize (0, nsid);
851 bool make_consistent = false;
853 /* Get file information. */
854 if (__builtin_expect (__fxstat64 (_STAT_VER, fd, &st) < 0, 0))
856 errstring = N_("cannot stat shared object");
857 call_lose_errno:
858 errval = errno;
859 call_lose:
860 lose (errval, fd, name, realname, l, errstring,
861 make_consistent ? r : NULL);
864 /* Look again to see if the real name matched another already loaded. */
865 for (l = GL(dl_ns)[nsid]._ns_loaded; l; l = l->l_next)
866 if (l->l_removed == 0 && l->l_ino == st.st_ino && l->l_dev == st.st_dev)
868 /* The object is already loaded.
869 Just bump its reference count and return it. */
870 __close (fd);
872 /* If the name is not in the list of names for this object add
873 it. */
874 free (realname);
875 add_name_to_object (l, name);
877 return l;
880 #ifdef SHARED
881 /* When loading into a namespace other than the base one we must
882 avoid loading ld.so since there can only be one copy. Ever. */
883 if (__builtin_expect (nsid != LM_ID_BASE, 0)
884 && ((st.st_ino == GL(dl_rtld_map).l_ino
885 && st.st_dev == GL(dl_rtld_map).l_dev)
886 || _dl_name_match_p (name, &GL(dl_rtld_map))))
888 /* This is indeed ld.so. Create a new link_map which refers to
889 the real one for almost everything. */
890 l = _dl_new_object (realname, name, l_type, loader, mode, nsid);
891 if (l == NULL)
892 goto fail_new;
894 /* Refer to the real descriptor. */
895 l->l_real = &GL(dl_rtld_map);
897 /* No need to bump the refcount of the real object, ld.so will
898 never be unloaded. */
899 __close (fd);
901 return l;
903 #endif
905 if (mode & RTLD_NOLOAD)
906 /* We are not supposed to load the object unless it is already
907 loaded. So return now. */
908 return NULL;
910 /* Print debugging message. */
911 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_FILES, 0))
912 _dl_debug_printf ("file=%s [%lu]; generating link map\n", name, nsid);
914 /* This is the ELF header. We read it in `open_verify'. */
915 header = (void *) fbp->buf;
917 #ifndef MAP_ANON
918 # define MAP_ANON 0
919 if (_dl_zerofd == -1)
921 _dl_zerofd = _dl_sysdep_open_zero_fill ();
922 if (_dl_zerofd == -1)
924 __close (fd);
925 _dl_signal_error (errno, NULL, NULL,
926 N_("cannot open zero fill device"));
929 #endif
931 /* Signal that we are going to add new objects. */
932 if (r->r_state == RT_CONSISTENT)
934 #ifdef SHARED
935 /* Auditing checkpoint: we are going to add new objects. */
936 if ((mode & __RTLD_AUDIT) == 0
937 && __builtin_expect (GLRO(dl_naudit) > 0, 0))
939 struct link_map *head = GL(dl_ns)[nsid]._ns_loaded;
940 /* Do not call the functions for any auditing object. */
941 if (head->l_auditing == 0)
943 struct audit_ifaces *afct = GLRO(dl_audit);
944 for (unsigned int cnt = 0; cnt < GLRO(dl_naudit); ++cnt)
946 if (afct->activity != NULL)
947 afct->activity (&head->l_audit[cnt].cookie, LA_ACT_ADD);
949 afct = afct->next;
953 #endif
955 /* Notify the debugger we have added some objects. We need to
956 call _dl_debug_initialize in a static program in case dynamic
957 linking has not been used before. */
958 r->r_state = RT_ADD;
959 _dl_debug_state ();
960 make_consistent = true;
962 else
963 assert (r->r_state == RT_ADD);
965 /* Enter the new object in the list of loaded objects. */
966 l = _dl_new_object (realname, name, l_type, loader, mode, nsid);
967 if (__builtin_expect (l == NULL, 0))
969 #ifdef SHARED
970 fail_new:
971 #endif
972 errstring = N_("cannot create shared object descriptor");
973 goto call_lose_errno;
976 /* Extract the remaining details we need from the ELF header
977 and then read in the program header table. */
978 l->l_entry = header->e_entry;
979 type = header->e_type;
980 l->l_phnum = header->e_phnum;
982 maplength = header->e_phnum * sizeof (ElfW(Phdr));
983 if (header->e_phoff + maplength <= (size_t) fbp->len)
984 phdr = (void *) (fbp->buf + header->e_phoff);
985 else
987 phdr = alloca (maplength);
988 __lseek (fd, header->e_phoff, SEEK_SET);
989 if ((size_t) __libc_read (fd, (void *) phdr, maplength) != maplength)
991 errstring = N_("cannot read file data");
992 goto call_lose_errno;
996 /* Presumed absent PT_GNU_STACK. */
997 uint_fast16_t stack_flags = PF_R|PF_W|PF_X;
1000 /* Scan the program header table, collecting its load commands. */
1001 struct loadcmd
1003 ElfW(Addr) mapstart, mapend, dataend, allocend;
1004 off_t mapoff;
1005 int prot;
1006 } loadcmds[l->l_phnum], *c;
1007 size_t nloadcmds = 0;
1008 bool has_holes = false;
1010 /* The struct is initialized to zero so this is not necessary:
1011 l->l_ld = 0;
1012 l->l_phdr = 0;
1013 l->l_addr = 0; */
1014 for (ph = phdr; ph < &phdr[l->l_phnum]; ++ph)
1015 switch (ph->p_type)
1017 /* These entries tell us where to find things once the file's
1018 segments are mapped in. We record the addresses it says
1019 verbatim, and later correct for the run-time load address. */
1020 case PT_DYNAMIC:
1021 l->l_ld = (void *) ph->p_vaddr;
1022 l->l_ldnum = ph->p_memsz / sizeof (ElfW(Dyn));
1023 break;
1025 case PT_PHDR:
1026 l->l_phdr = (void *) ph->p_vaddr;
1027 break;
1029 case PT_LOAD:
1030 /* A load command tells us to map in part of the file.
1031 We record the load commands and process them all later. */
1032 if (__builtin_expect ((ph->p_align & (GLRO(dl_pagesize) - 1)) != 0,
1035 errstring = N_("ELF load command alignment not page-aligned");
1036 goto call_lose;
1038 if (__builtin_expect (((ph->p_vaddr - ph->p_offset)
1039 & (ph->p_align - 1)) != 0, 0))
1041 errstring
1042 = N_("ELF load command address/offset not properly aligned");
1043 goto call_lose;
1046 c = &loadcmds[nloadcmds++];
1047 c->mapstart = ph->p_vaddr & ~(GLRO(dl_pagesize) - 1);
1048 c->mapend = ((ph->p_vaddr + ph->p_filesz + GLRO(dl_pagesize) - 1)
1049 & ~(GLRO(dl_pagesize) - 1));
1050 c->dataend = ph->p_vaddr + ph->p_filesz;
1051 c->allocend = ph->p_vaddr + ph->p_memsz;
1052 c->mapoff = ph->p_offset & ~(GLRO(dl_pagesize) - 1);
1054 /* Determine whether there is a gap between the last segment
1055 and this one. */
1056 if (nloadcmds > 1 && c[-1].mapend != c->mapstart)
1057 has_holes = true;
1059 /* Optimize a common case. */
1060 #if (PF_R | PF_W | PF_X) == 7 && (PROT_READ | PROT_WRITE | PROT_EXEC) == 7
1061 c->prot = (PF_TO_PROT
1062 >> ((ph->p_flags & (PF_R | PF_W | PF_X)) * 4)) & 0xf;
1063 #else
1064 c->prot = 0;
1065 if (ph->p_flags & PF_R)
1066 c->prot |= PROT_READ;
1067 if (ph->p_flags & PF_W)
1068 c->prot |= PROT_WRITE;
1069 if (ph->p_flags & PF_X)
1070 c->prot |= PROT_EXEC;
1071 #endif
1072 break;
1074 case PT_TLS:
1075 if (ph->p_memsz == 0)
1076 /* Nothing to do for an empty segment. */
1077 break;
1079 l->l_tls_blocksize = ph->p_memsz;
1080 l->l_tls_align = ph->p_align;
1081 if (ph->p_align == 0)
1082 l->l_tls_firstbyte_offset = 0;
1083 else
1084 l->l_tls_firstbyte_offset = ph->p_vaddr & (ph->p_align - 1);
1085 l->l_tls_initimage_size = ph->p_filesz;
1086 /* Since we don't know the load address yet only store the
1087 offset. We will adjust it later. */
1088 l->l_tls_initimage = (void *) ph->p_vaddr;
1090 /* If not loading the initial set of shared libraries,
1091 check whether we should permit loading a TLS segment. */
1092 if (__builtin_expect (l->l_type == lt_library, 1)
1093 /* If GL(dl_tls_dtv_slotinfo_list) == NULL, then rtld.c did
1094 not set up TLS data structures, so don't use them now. */
1095 || __builtin_expect (GL(dl_tls_dtv_slotinfo_list) != NULL, 1))
1097 /* Assign the next available module ID. */
1098 l->l_tls_modid = _dl_next_tls_modid ();
1099 break;
1102 #ifdef SHARED
1103 if (l->l_prev == NULL || (mode & __RTLD_AUDIT) != 0)
1104 /* We are loading the executable itself when the dynamic linker
1105 was executed directly. The setup will happen later. */
1106 break;
1108 /* In a static binary there is no way to tell if we dynamically
1109 loaded libpthread. */
1110 if (GL(dl_error_catch_tsd) == &_dl_initial_error_catch_tsd)
1111 #endif
1113 /* We have not yet loaded libpthread.
1114 We can do the TLS setup right now! */
1116 void *tcb;
1118 /* The first call allocates TLS bookkeeping data structures.
1119 Then we allocate the TCB for the initial thread. */
1120 if (__builtin_expect (_dl_tls_setup (), 0)
1121 || __builtin_expect ((tcb = _dl_allocate_tls (NULL)) == NULL,
1124 errval = ENOMEM;
1125 errstring = N_("\
1126 cannot allocate TLS data structures for initial thread");
1127 goto call_lose;
1130 /* Now we install the TCB in the thread register. */
1131 errstring = TLS_INIT_TP (tcb, 0);
1132 if (__builtin_expect (errstring == NULL, 1))
1134 /* Now we are all good. */
1135 l->l_tls_modid = ++GL(dl_tls_max_dtv_idx);
1136 break;
1139 /* The kernel is too old or somesuch. */
1140 errval = 0;
1141 _dl_deallocate_tls (tcb, 1);
1142 goto call_lose;
1145 /* Uh-oh, the binary expects TLS support but we cannot
1146 provide it. */
1147 errval = 0;
1148 errstring = N_("cannot handle TLS data");
1149 goto call_lose;
1150 break;
1152 case PT_GNU_STACK:
1153 stack_flags = ph->p_flags;
1154 break;
1156 case PT_GNU_RELRO:
1157 l->l_relro_addr = ph->p_vaddr;
1158 l->l_relro_size = ph->p_memsz;
1159 break;
1162 if (__builtin_expect (nloadcmds == 0, 0))
1164 /* This only happens for a bogus object that will be caught with
1165 another error below. But we don't want to go through the
1166 calculations below using NLOADCMDS - 1. */
1167 errstring = N_("object file has no loadable segments");
1168 goto call_lose;
1171 /* Now process the load commands and map segments into memory. */
1172 c = loadcmds;
1174 /* Length of the sections to be loaded. */
1175 maplength = loadcmds[nloadcmds - 1].allocend - c->mapstart;
1177 if (__builtin_expect (type, ET_DYN) == ET_DYN)
1179 /* This is a position-independent shared object. We can let the
1180 kernel map it anywhere it likes, but we must have space for all
1181 the segments in their specified positions relative to the first.
1182 So we map the first segment without MAP_FIXED, but with its
1183 extent increased to cover all the segments. Then we remove
1184 access from excess portion, and there is known sufficient space
1185 there to remap from the later segments.
1187 As a refinement, sometimes we have an address that we would
1188 prefer to map such objects at; but this is only a preference,
1189 the OS can do whatever it likes. */
1190 ElfW(Addr) mappref;
1191 mappref = (ELF_PREFERRED_ADDRESS (loader, maplength,
1192 c->mapstart & GLRO(dl_use_load_bias))
1193 - MAP_BASE_ADDR (l));
1195 /* Remember which part of the address space this object uses. */
1196 l->l_map_start = (ElfW(Addr)) __mmap ((void *) mappref, maplength,
1197 c->prot,
1198 MAP_COPY|MAP_FILE,
1199 fd, c->mapoff);
1200 if (__builtin_expect ((void *) l->l_map_start == MAP_FAILED, 0))
1202 map_error:
1203 errstring = N_("failed to map segment from shared object");
1204 goto call_lose_errno;
1207 l->l_map_end = l->l_map_start + maplength;
1208 l->l_addr = l->l_map_start - c->mapstart;
1210 if (has_holes)
1211 /* Change protection on the excess portion to disallow all access;
1212 the portions we do not remap later will be inaccessible as if
1213 unallocated. Then jump into the normal segment-mapping loop to
1214 handle the portion of the segment past the end of the file
1215 mapping. */
1216 __mprotect ((caddr_t) (l->l_addr + c->mapend),
1217 loadcmds[nloadcmds - 1].mapstart - c->mapend,
1218 PROT_NONE);
1220 l->l_contiguous = 1;
1222 goto postmap;
1225 /* This object is loaded at a fixed address. This must never
1226 happen for objects loaded with dlopen(). */
1227 if (__builtin_expect ((mode & __RTLD_OPENEXEC) == 0, 0))
1229 errstring = N_("cannot dynamically load executable");
1230 goto call_lose;
1233 /* Notify ELF_PREFERRED_ADDRESS that we have to load this one
1234 fixed. */
1235 ELF_FIXED_ADDRESS (loader, c->mapstart);
1238 /* Remember which part of the address space this object uses. */
1239 l->l_map_start = c->mapstart + l->l_addr;
1240 l->l_map_end = l->l_map_start + maplength;
1241 l->l_contiguous = !has_holes;
1243 while (c < &loadcmds[nloadcmds])
1245 if (c->mapend > c->mapstart
1246 /* Map the segment contents from the file. */
1247 && (__mmap ((void *) (l->l_addr + c->mapstart),
1248 c->mapend - c->mapstart, c->prot,
1249 MAP_FIXED|MAP_COPY|MAP_FILE,
1250 fd, c->mapoff)
1251 == MAP_FAILED))
1252 goto map_error;
1254 postmap:
1255 if (c->prot & PROT_EXEC)
1256 l->l_text_end = l->l_addr + c->mapend;
1258 if (l->l_phdr == 0
1259 && (ElfW(Off)) c->mapoff <= header->e_phoff
1260 && ((size_t) (c->mapend - c->mapstart + c->mapoff)
1261 >= header->e_phoff + header->e_phnum * sizeof (ElfW(Phdr))))
1262 /* Found the program header in this segment. */
1263 l->l_phdr = (void *) (c->mapstart + header->e_phoff - c->mapoff);
1265 if (c->allocend > c->dataend)
1267 /* Extra zero pages should appear at the end of this segment,
1268 after the data mapped from the file. */
1269 ElfW(Addr) zero, zeroend, zeropage;
1271 zero = l->l_addr + c->dataend;
1272 zeroend = l->l_addr + c->allocend;
1273 zeropage = ((zero + GLRO(dl_pagesize) - 1)
1274 & ~(GLRO(dl_pagesize) - 1));
1276 if (zeroend < zeropage)
1277 /* All the extra data is in the last page of the segment.
1278 We can just zero it. */
1279 zeropage = zeroend;
1281 if (zeropage > zero)
1283 /* Zero the final part of the last page of the segment. */
1284 if (__builtin_expect ((c->prot & PROT_WRITE) == 0, 0))
1286 /* Dag nab it. */
1287 if (__mprotect ((caddr_t) (zero
1288 & ~(GLRO(dl_pagesize) - 1)),
1289 GLRO(dl_pagesize), c->prot|PROT_WRITE) < 0)
1291 errstring = N_("cannot change memory protections");
1292 goto call_lose_errno;
1295 memset ((void *) zero, '\0', zeropage - zero);
1296 if (__builtin_expect ((c->prot & PROT_WRITE) == 0, 0))
1297 __mprotect ((caddr_t) (zero & ~(GLRO(dl_pagesize) - 1)),
1298 GLRO(dl_pagesize), c->prot);
1301 if (zeroend > zeropage)
1303 /* Map the remaining zero pages in from the zero fill FD. */
1304 caddr_t mapat;
1305 mapat = __mmap ((caddr_t) zeropage, zeroend - zeropage,
1306 c->prot, MAP_ANON|MAP_PRIVATE|MAP_FIXED,
1307 -1, 0);
1308 if (__builtin_expect (mapat == MAP_FAILED, 0))
1310 errstring = N_("cannot map zero-fill pages");
1311 goto call_lose_errno;
1316 ++c;
1320 if (l->l_ld == 0)
1322 if (__builtin_expect (type == ET_DYN, 0))
1324 errstring = N_("object file has no dynamic section");
1325 goto call_lose;
1328 else
1329 l->l_ld = (ElfW(Dyn) *) ((ElfW(Addr)) l->l_ld + l->l_addr);
1331 elf_get_dynamic_info (l, NULL);
1333 /* Make sure we are not dlopen'ing an object that has the
1334 DF_1_NOOPEN flag set. */
1335 if (__builtin_expect (l->l_flags_1 & DF_1_NOOPEN, 0)
1336 && (mode & __RTLD_DLOPEN))
1338 /* We are not supposed to load this object. Free all resources. */
1339 __munmap ((void *) l->l_map_start, l->l_map_end - l->l_map_start);
1341 if (!l->l_libname->dont_free)
1342 free (l->l_libname);
1344 if (l->l_phdr_allocated)
1345 free ((void *) l->l_phdr);
1347 errstring = N_("shared object cannot be dlopen()ed");
1348 goto call_lose;
1351 if (l->l_phdr == NULL)
1353 /* The program header is not contained in any of the segments.
1354 We have to allocate memory ourself and copy it over from out
1355 temporary place. */
1356 ElfW(Phdr) *newp = (ElfW(Phdr) *) malloc (header->e_phnum
1357 * sizeof (ElfW(Phdr)));
1358 if (newp == NULL)
1360 errstring = N_("cannot allocate memory for program header");
1361 goto call_lose_errno;
1364 l->l_phdr = memcpy (newp, phdr,
1365 (header->e_phnum * sizeof (ElfW(Phdr))));
1366 l->l_phdr_allocated = 1;
1368 else
1369 /* Adjust the PT_PHDR value by the runtime load address. */
1370 l->l_phdr = (ElfW(Phdr) *) ((ElfW(Addr)) l->l_phdr + l->l_addr);
1372 if (__builtin_expect ((stack_flags &~ GL(dl_stack_flags)) & PF_X, 0))
1374 if (__builtin_expect (__check_caller (RETURN_ADDRESS (0), allow_ldso),
1375 0) != 0)
1377 errstring = N_("invalid caller");
1378 goto call_lose;
1381 /* The stack is presently not executable, but this module
1382 requires that it be executable. We must change the
1383 protection of the variable which contains the flags used in
1384 the mprotect calls. */
1385 #ifdef SHARED
1386 if ((mode & (__RTLD_DLOPEN | __RTLD_AUDIT)) == __RTLD_DLOPEN)
1388 const uintptr_t p = (uintptr_t) &__stack_prot & -GLRO(dl_pagesize);
1389 const size_t s = (uintptr_t) (&__stack_prot + 1) - p;
1391 struct link_map *const m = &GL(dl_rtld_map);
1392 const uintptr_t relro_end = ((m->l_addr + m->l_relro_addr
1393 + m->l_relro_size)
1394 & -GLRO(dl_pagesize));
1395 if (__builtin_expect (p + s <= relro_end, 1))
1397 /* The variable lies in the region protected by RELRO. */
1398 __mprotect ((void *) p, s, PROT_READ|PROT_WRITE);
1399 __stack_prot |= PROT_READ|PROT_WRITE|PROT_EXEC;
1400 __mprotect ((void *) p, s, PROT_READ);
1402 else
1403 __stack_prot |= PROT_READ|PROT_WRITE|PROT_EXEC;
1405 else
1406 #endif
1407 __stack_prot |= PROT_READ|PROT_WRITE|PROT_EXEC;
1409 #ifdef check_consistency
1410 check_consistency ();
1411 #endif
1413 errval = (*GL(dl_make_stack_executable_hook)) (stack_endp);
1414 if (errval)
1416 errstring = N_("\
1417 cannot enable executable stack as shared object requires");
1418 goto call_lose;
1422 /* Adjust the address of the TLS initialization image. */
1423 if (l->l_tls_initimage != NULL)
1424 l->l_tls_initimage = (char *) l->l_tls_initimage + l->l_addr;
1426 /* We are done mapping in the file. We no longer need the descriptor. */
1427 if (__builtin_expect (__close (fd) != 0, 0))
1429 errstring = N_("cannot close file descriptor");
1430 goto call_lose_errno;
1432 /* Signal that we closed the file. */
1433 fd = -1;
1435 if (l->l_type == lt_library && type == ET_EXEC)
1436 l->l_type = lt_executable;
1438 l->l_entry += l->l_addr;
1440 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_FILES, 0))
1441 _dl_debug_printf ("\
1442 dynamic: 0x%0*lx base: 0x%0*lx size: 0x%0*Zx\n\
1443 entry: 0x%0*lx phdr: 0x%0*lx phnum: %*u\n\n",
1444 (int) sizeof (void *) * 2,
1445 (unsigned long int) l->l_ld,
1446 (int) sizeof (void *) * 2,
1447 (unsigned long int) l->l_addr,
1448 (int) sizeof (void *) * 2, maplength,
1449 (int) sizeof (void *) * 2,
1450 (unsigned long int) l->l_entry,
1451 (int) sizeof (void *) * 2,
1452 (unsigned long int) l->l_phdr,
1453 (int) sizeof (void *) * 2, l->l_phnum);
1455 /* Set up the symbol hash table. */
1456 _dl_setup_hash (l);
1458 /* If this object has DT_SYMBOLIC set modify now its scope. We don't
1459 have to do this for the main map. */
1460 if ((mode & RTLD_DEEPBIND) == 0
1461 && __builtin_expect (l->l_info[DT_SYMBOLIC] != NULL, 0)
1462 && &l->l_searchlist != l->l_scope[0])
1464 /* Create an appropriate searchlist. It contains only this map.
1465 This is the definition of DT_SYMBOLIC in SysVr4. */
1466 l->l_symbolic_searchlist.r_list[0] = l;
1467 l->l_symbolic_searchlist.r_nlist = 1;
1469 /* Now move the existing entries one back. */
1470 memmove (&l->l_scope[1], &l->l_scope[0],
1471 (l->l_scope_max - 1) * sizeof (l->l_scope[0]));
1473 /* Now add the new entry. */
1474 l->l_scope[0] = &l->l_symbolic_searchlist;
1477 /* Remember whether this object must be initialized first. */
1478 if (l->l_flags_1 & DF_1_INITFIRST)
1479 GL(dl_initfirst) = l;
1481 /* Finally the file information. */
1482 l->l_dev = st.st_dev;
1483 l->l_ino = st.st_ino;
1485 /* When we profile the SONAME might be needed for something else but
1486 loading. Add it right away. */
1487 if (__builtin_expect (GLRO(dl_profile) != NULL, 0)
1488 && l->l_info[DT_SONAME] != NULL)
1489 add_name_to_object (l, ((const char *) D_PTR (l, l_info[DT_STRTAB])
1490 + l->l_info[DT_SONAME]->d_un.d_val));
1492 #ifdef SHARED
1493 /* Auditing checkpoint: we have a new object. */
1494 if (__builtin_expect (GLRO(dl_naudit) > 0, 0)
1495 && !GL(dl_ns)[l->l_ns]._ns_loaded->l_auditing)
1497 struct audit_ifaces *afct = GLRO(dl_audit);
1498 for (unsigned int cnt = 0; cnt < GLRO(dl_naudit); ++cnt)
1500 if (afct->objopen != NULL)
1502 l->l_audit[cnt].bindflags
1503 = afct->objopen (l, nsid, &l->l_audit[cnt].cookie);
1505 l->l_audit_any_plt |= l->l_audit[cnt].bindflags != 0;
1508 afct = afct->next;
1511 #endif
1513 return l;
1516 /* Print search path. */
1517 static void
1518 print_search_path (struct r_search_path_elem **list,
1519 const char *what, const char *name)
1521 char buf[max_dirnamelen + max_capstrlen];
1522 int first = 1;
1524 _dl_debug_printf (" search path=");
1526 while (*list != NULL && (*list)->what == what) /* Yes, ==. */
1528 char *endp = __mempcpy (buf, (*list)->dirname, (*list)->dirnamelen);
1529 size_t cnt;
1531 for (cnt = 0; cnt < ncapstr; ++cnt)
1532 if ((*list)->status[cnt] != nonexisting)
1534 char *cp = __mempcpy (endp, capstr[cnt].str, capstr[cnt].len);
1535 if (cp == buf || (cp == buf + 1 && buf[0] == '/'))
1536 cp[0] = '\0';
1537 else
1538 cp[-1] = '\0';
1540 _dl_debug_printf_c (first ? "%s" : ":%s", buf);
1541 first = 0;
1544 ++list;
1547 if (name != NULL)
1548 _dl_debug_printf_c ("\t\t(%s from file %s)\n", what,
1549 name[0] ? name : rtld_progname);
1550 else
1551 _dl_debug_printf_c ("\t\t(%s)\n", what);
1554 /* Open a file and verify it is an ELF file for this architecture. We
1555 ignore only ELF files for other architectures. Non-ELF files and
1556 ELF files with different header information cause fatal errors since
1557 this could mean there is something wrong in the installation and the
1558 user might want to know about this. */
1559 static int
1560 open_verify (const char *name, struct filebuf *fbp, struct link_map *loader,
1561 int whatcode, bool *found_other_class, bool free_name)
1563 /* This is the expected ELF header. */
1564 #define ELF32_CLASS ELFCLASS32
1565 #define ELF64_CLASS ELFCLASS64
1566 #ifndef VALID_ELF_HEADER
1567 # define VALID_ELF_HEADER(hdr,exp,size) (memcmp (hdr, exp, size) == 0)
1568 # define VALID_ELF_OSABI(osabi) (osabi == ELFOSABI_SYSV)
1569 # define VALID_ELF_ABIVERSION(ver) (ver == 0)
1570 #elif defined MORE_ELF_HEADER_DATA
1571 MORE_ELF_HEADER_DATA;
1572 #endif
1573 static const unsigned char expected[EI_PAD] =
1575 [EI_MAG0] = ELFMAG0,
1576 [EI_MAG1] = ELFMAG1,
1577 [EI_MAG2] = ELFMAG2,
1578 [EI_MAG3] = ELFMAG3,
1579 [EI_CLASS] = ELFW(CLASS),
1580 [EI_DATA] = byteorder,
1581 [EI_VERSION] = EV_CURRENT,
1582 [EI_OSABI] = ELFOSABI_SYSV,
1583 [EI_ABIVERSION] = 0
1585 static const struct
1587 ElfW(Word) vendorlen;
1588 ElfW(Word) datalen;
1589 ElfW(Word) type;
1590 char vendor[4];
1591 } expected_note = { 4, 16, 1, "GNU" };
1592 /* Initialize it to make the compiler happy. */
1593 const char *errstring = NULL;
1594 int errval = 0;
1596 #ifdef SHARED
1597 /* Give the auditing libraries a chance. */
1598 if (__builtin_expect (GLRO(dl_naudit) > 0, 0) && whatcode != 0
1599 && loader->l_auditing == 0)
1601 struct audit_ifaces *afct = GLRO(dl_audit);
1602 for (unsigned int cnt = 0; cnt < GLRO(dl_naudit); ++cnt)
1604 if (afct->objsearch != NULL)
1606 name = afct->objsearch (name, &loader->l_audit[cnt].cookie,
1607 whatcode);
1608 if (name == NULL)
1609 /* Ignore the path. */
1610 return -1;
1613 afct = afct->next;
1616 #endif
1618 /* Open the file. We always open files read-only. */
1619 int fd = __open (name, O_RDONLY);
1620 if (fd != -1)
1622 ElfW(Ehdr) *ehdr;
1623 ElfW(Phdr) *phdr, *ph;
1624 ElfW(Word) *abi_note;
1625 unsigned int osversion;
1626 size_t maplength;
1628 /* We successfully openened the file. Now verify it is a file
1629 we can use. */
1630 __set_errno (0);
1631 fbp->len = __libc_read (fd, fbp->buf, sizeof (fbp->buf));
1633 /* This is where the ELF header is loaded. */
1634 assert (sizeof (fbp->buf) > sizeof (ElfW(Ehdr)));
1635 ehdr = (ElfW(Ehdr) *) fbp->buf;
1637 /* Now run the tests. */
1638 if (__builtin_expect (fbp->len < (ssize_t) sizeof (ElfW(Ehdr)), 0))
1640 errval = errno;
1641 errstring = (errval == 0
1642 ? N_("file too short") : N_("cannot read file data"));
1643 call_lose:
1644 if (free_name)
1646 char *realname = (char *) name;
1647 name = strdupa (realname);
1648 free (realname);
1650 lose (errval, fd, name, NULL, NULL, errstring, NULL);
1653 /* See whether the ELF header is what we expect. */
1654 if (__builtin_expect (! VALID_ELF_HEADER (ehdr->e_ident, expected,
1655 EI_PAD), 0))
1657 /* Something is wrong. */
1658 const Elf32_Word *magp = (const void *) ehdr->e_ident;
1659 if (*magp !=
1660 #if BYTE_ORDER == LITTLE_ENDIAN
1661 ((ELFMAG0 << (EI_MAG0 * 8)) |
1662 (ELFMAG1 << (EI_MAG1 * 8)) |
1663 (ELFMAG2 << (EI_MAG2 * 8)) |
1664 (ELFMAG3 << (EI_MAG3 * 8)))
1665 #else
1666 ((ELFMAG0 << (EI_MAG3 * 8)) |
1667 (ELFMAG1 << (EI_MAG2 * 8)) |
1668 (ELFMAG2 << (EI_MAG1 * 8)) |
1669 (ELFMAG3 << (EI_MAG0 * 8)))
1670 #endif
1672 errstring = N_("invalid ELF header");
1673 else if (ehdr->e_ident[EI_CLASS] != ELFW(CLASS))
1675 /* This is not a fatal error. On architectures where
1676 32-bit and 64-bit binaries can be run this might
1677 happen. */
1678 *found_other_class = true;
1679 goto close_and_out;
1681 else if (ehdr->e_ident[EI_DATA] != byteorder)
1683 if (BYTE_ORDER == BIG_ENDIAN)
1684 errstring = N_("ELF file data encoding not big-endian");
1685 else
1686 errstring = N_("ELF file data encoding not little-endian");
1688 else if (ehdr->e_ident[EI_VERSION] != EV_CURRENT)
1689 errstring
1690 = N_("ELF file version ident does not match current one");
1691 /* XXX We should be able so set system specific versions which are
1692 allowed here. */
1693 else if (!VALID_ELF_OSABI (ehdr->e_ident[EI_OSABI]))
1694 errstring = N_("ELF file OS ABI invalid");
1695 else if (!VALID_ELF_ABIVERSION (ehdr->e_ident[EI_ABIVERSION]))
1696 errstring = N_("ELF file ABI version invalid");
1697 else
1698 /* Otherwise we don't know what went wrong. */
1699 errstring = N_("internal error");
1701 goto call_lose;
1704 if (__builtin_expect (ehdr->e_version, EV_CURRENT) != EV_CURRENT)
1706 errstring = N_("ELF file version does not match current one");
1707 goto call_lose;
1709 if (! __builtin_expect (elf_machine_matches_host (ehdr), 1))
1710 goto close_and_out;
1711 else if (__builtin_expect (ehdr->e_type, ET_DYN) != ET_DYN
1712 && __builtin_expect (ehdr->e_type, ET_EXEC) != ET_EXEC)
1714 errstring = N_("only ET_DYN and ET_EXEC can be loaded");
1715 goto call_lose;
1717 else if (__builtin_expect (ehdr->e_phentsize, sizeof (ElfW(Phdr)))
1718 != sizeof (ElfW(Phdr)))
1720 errstring = N_("ELF file's phentsize not the expected size");
1721 goto call_lose;
1724 maplength = ehdr->e_phnum * sizeof (ElfW(Phdr));
1725 if (ehdr->e_phoff + maplength <= (size_t) fbp->len)
1726 phdr = (void *) (fbp->buf + ehdr->e_phoff);
1727 else
1729 phdr = alloca (maplength);
1730 __lseek (fd, ehdr->e_phoff, SEEK_SET);
1731 if ((size_t) __libc_read (fd, (void *) phdr, maplength) != maplength)
1733 read_error:
1734 errval = errno;
1735 errstring = N_("cannot read file data");
1736 goto call_lose;
1740 /* Check .note.ABI-tag if present. */
1741 for (ph = phdr; ph < &phdr[ehdr->e_phnum]; ++ph)
1742 if (ph->p_type == PT_NOTE && ph->p_filesz >= 32 && ph->p_align >= 4)
1744 ElfW(Addr) size = ph->p_filesz;
1746 if (ph->p_offset + size <= (size_t) fbp->len)
1747 abi_note = (void *) (fbp->buf + ph->p_offset);
1748 else
1750 abi_note = alloca (size);
1751 __lseek (fd, ph->p_offset, SEEK_SET);
1752 if (__libc_read (fd, (void *) abi_note, size) != size)
1753 goto read_error;
1756 while (memcmp (abi_note, &expected_note, sizeof (expected_note)))
1758 #define ROUND(len) (((len) + sizeof (ElfW(Word)) - 1) & -sizeof (ElfW(Word)))
1759 ElfW(Addr) note_size = 3 * sizeof (ElfW(Word))
1760 + ROUND (abi_note[0])
1761 + ROUND (abi_note[1]);
1763 if (size - 32 < note_size)
1765 size = 0;
1766 break;
1768 size -= note_size;
1769 abi_note = (void *) abi_note + note_size;
1772 if (size == 0)
1773 continue;
1775 osversion = (abi_note[5] & 0xff) * 65536
1776 + (abi_note[6] & 0xff) * 256
1777 + (abi_note[7] & 0xff);
1778 if (abi_note[4] != __ABI_TAG_OS
1779 || (GLRO(dl_osversion) && GLRO(dl_osversion) < osversion))
1781 close_and_out:
1782 __close (fd);
1783 __set_errno (ENOENT);
1784 fd = -1;
1787 break;
1791 return fd;
1794 /* Try to open NAME in one of the directories in *DIRSP.
1795 Return the fd, or -1. If successful, fill in *REALNAME
1796 with the malloc'd full directory name. If it turns out
1797 that none of the directories in *DIRSP exists, *DIRSP is
1798 replaced with (void *) -1, and the old value is free()d
1799 if MAY_FREE_DIRS is true. */
1801 static int
1802 open_path (const char *name, size_t namelen, int preloaded,
1803 struct r_search_path_struct *sps, char **realname,
1804 struct filebuf *fbp, struct link_map *loader, int whatcode,
1805 bool *found_other_class)
1807 struct r_search_path_elem **dirs = sps->dirs;
1808 char *buf;
1809 int fd = -1;
1810 const char *current_what = NULL;
1811 int any = 0;
1813 if (__builtin_expect (dirs == NULL, 0))
1814 /* We're called before _dl_init_paths when loading the main executable
1815 given on the command line when rtld is run directly. */
1816 return -1;
1818 buf = alloca (max_dirnamelen + max_capstrlen + namelen);
1821 struct r_search_path_elem *this_dir = *dirs;
1822 size_t buflen = 0;
1823 size_t cnt;
1824 char *edp;
1825 int here_any = 0;
1826 int err;
1828 /* If we are debugging the search for libraries print the path
1829 now if it hasn't happened now. */
1830 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_LIBS, 0)
1831 && current_what != this_dir->what)
1833 current_what = this_dir->what;
1834 print_search_path (dirs, current_what, this_dir->where);
1837 edp = (char *) __mempcpy (buf, this_dir->dirname, this_dir->dirnamelen);
1838 for (cnt = 0; fd == -1 && cnt < ncapstr; ++cnt)
1840 /* Skip this directory if we know it does not exist. */
1841 if (this_dir->status[cnt] == nonexisting)
1842 continue;
1844 buflen =
1845 ((char *) __mempcpy (__mempcpy (edp, capstr[cnt].str,
1846 capstr[cnt].len),
1847 name, namelen)
1848 - buf);
1850 /* Print name we try if this is wanted. */
1851 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_LIBS, 0))
1852 _dl_debug_printf (" trying file=%s\n", buf);
1854 fd = open_verify (buf, fbp, loader, whatcode, found_other_class,
1855 false);
1856 if (this_dir->status[cnt] == unknown)
1858 if (fd != -1)
1859 this_dir->status[cnt] = existing;
1860 /* Do not update the directory information when loading
1861 auditing code. We must try to disturb the program as
1862 little as possible. */
1863 else if (loader == NULL
1864 || GL(dl_ns)[loader->l_ns]._ns_loaded->l_auditing == 0)
1866 /* We failed to open machine dependent library. Let's
1867 test whether there is any directory at all. */
1868 struct stat64 st;
1870 buf[buflen - namelen - 1] = '\0';
1872 if (__xstat64 (_STAT_VER, buf, &st) != 0
1873 || ! S_ISDIR (st.st_mode))
1874 /* The directory does not exist or it is no directory. */
1875 this_dir->status[cnt] = nonexisting;
1876 else
1877 this_dir->status[cnt] = existing;
1881 /* Remember whether we found any existing directory. */
1882 here_any |= this_dir->status[cnt] != nonexisting;
1884 if (fd != -1 && __builtin_expect (preloaded, 0)
1885 && INTUSE(__libc_enable_secure))
1887 /* This is an extra security effort to make sure nobody can
1888 preload broken shared objects which are in the trusted
1889 directories and so exploit the bugs. */
1890 struct stat64 st;
1892 if (__fxstat64 (_STAT_VER, fd, &st) != 0
1893 || (st.st_mode & S_ISUID) == 0)
1895 /* The shared object cannot be tested for being SUID
1896 or this bit is not set. In this case we must not
1897 use this object. */
1898 __close (fd);
1899 fd = -1;
1900 /* We simply ignore the file, signal this by setting
1901 the error value which would have been set by `open'. */
1902 errno = ENOENT;
1907 if (fd != -1)
1909 *realname = (char *) malloc (buflen);
1910 if (*realname != NULL)
1912 memcpy (*realname, buf, buflen);
1913 return fd;
1915 else
1917 /* No memory for the name, we certainly won't be able
1918 to load and link it. */
1919 __close (fd);
1920 return -1;
1923 if (here_any && (err = errno) != ENOENT && err != EACCES)
1924 /* The file exists and is readable, but something went wrong. */
1925 return -1;
1927 /* Remember whether we found anything. */
1928 any |= here_any;
1930 while (*++dirs != NULL);
1932 /* Remove the whole path if none of the directories exists. */
1933 if (__builtin_expect (! any, 0))
1935 /* Paths which were allocated using the minimal malloc() in ld.so
1936 must not be freed using the general free() in libc. */
1937 if (sps->malloced)
1938 free (sps->dirs);
1940 /* rtld_search_dirs is attribute_relro, therefore avoid writing
1941 into it. */
1942 if (sps != &rtld_search_dirs)
1943 sps->dirs = (void *) -1;
1946 return -1;
1949 /* Map in the shared object file NAME. */
1951 struct link_map *
1952 internal_function
1953 _dl_map_object (struct link_map *loader, const char *name, int preloaded,
1954 int type, int trace_mode, int mode, Lmid_t nsid)
1956 int fd;
1957 char *realname;
1958 char *name_copy;
1959 struct link_map *l;
1960 struct filebuf fb;
1962 assert (nsid >= 0);
1963 assert (nsid < GL(dl_nns));
1965 /* Look for this name among those already loaded. */
1966 for (l = GL(dl_ns)[nsid]._ns_loaded; l; l = l->l_next)
1968 /* If the requested name matches the soname of a loaded object,
1969 use that object. Elide this check for names that have not
1970 yet been opened. */
1971 if (__builtin_expect (l->l_faked, 0) != 0
1972 || __builtin_expect (l->l_removed, 0) != 0)
1973 continue;
1974 if (!_dl_name_match_p (name, l))
1976 const char *soname;
1978 if (__builtin_expect (l->l_soname_added, 1)
1979 || l->l_info[DT_SONAME] == NULL)
1980 continue;
1982 soname = ((const char *) D_PTR (l, l_info[DT_STRTAB])
1983 + l->l_info[DT_SONAME]->d_un.d_val);
1984 if (strcmp (name, soname) != 0)
1985 continue;
1987 /* We have a match on a new name -- cache it. */
1988 add_name_to_object (l, soname);
1989 l->l_soname_added = 1;
1992 /* We have a match. */
1993 return l;
1996 /* Display information if we are debugging. */
1997 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_FILES, 0)
1998 && loader != NULL)
1999 _dl_debug_printf ("\nfile=%s [%lu]; needed by %s [%lu]\n", name, nsid,
2000 loader->l_name[0]
2001 ? loader->l_name : rtld_progname, loader->l_ns);
2003 #ifdef SHARED
2004 /* Give the auditing libraries a chance to change the name before we
2005 try anything. */
2006 if (__builtin_expect (GLRO(dl_naudit) > 0, 0)
2007 && (loader == NULL || loader->l_auditing == 0))
2009 struct audit_ifaces *afct = GLRO(dl_audit);
2010 for (unsigned int cnt = 0; cnt < GLRO(dl_naudit); ++cnt)
2012 if (afct->objsearch != NULL)
2014 name = afct->objsearch (name, &loader->l_audit[cnt].cookie,
2015 LA_SER_ORIG);
2016 if (name == NULL)
2018 /* Do not try anything further. */
2019 fd = -1;
2020 goto no_file;
2024 afct = afct->next;
2027 #endif
2029 /* Will be true if we found a DSO which is of the other ELF class. */
2030 bool found_other_class = false;
2032 if (strchr (name, '/') == NULL)
2034 /* Search for NAME in several places. */
2036 size_t namelen = strlen (name) + 1;
2038 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_LIBS, 0))
2039 _dl_debug_printf ("find library=%s [%lu]; searching\n", name, nsid);
2041 fd = -1;
2043 /* When the object has the RUNPATH information we don't use any
2044 RPATHs. */
2045 if (loader == NULL || loader->l_info[DT_RUNPATH] == NULL)
2047 /* This is the executable's map (if there is one). Make sure that
2048 we do not look at it twice. */
2049 struct link_map *main_map = GL(dl_ns)[LM_ID_BASE]._ns_loaded;
2050 bool did_main_map = false;
2052 /* First try the DT_RPATH of the dependent object that caused NAME
2053 to be loaded. Then that object's dependent, and on up. */
2054 for (l = loader; l; l = l->l_loader)
2055 if (cache_rpath (l, &l->l_rpath_dirs, DT_RPATH, "RPATH"))
2057 fd = open_path (name, namelen, preloaded, &l->l_rpath_dirs,
2058 &realname, &fb, loader, LA_SER_RUNPATH,
2059 &found_other_class);
2060 if (fd != -1)
2061 break;
2063 did_main_map |= l == main_map;
2066 /* If dynamically linked, try the DT_RPATH of the executable
2067 itself. NB: we do this for lookups in any namespace. */
2068 if (fd == -1 && !did_main_map
2069 && main_map != NULL && main_map->l_type != lt_loaded
2070 && cache_rpath (main_map, &main_map->l_rpath_dirs, DT_RPATH,
2071 "RPATH"))
2072 fd = open_path (name, namelen, preloaded, &main_map->l_rpath_dirs,
2073 &realname, &fb, loader ?: main_map, LA_SER_RUNPATH,
2074 &found_other_class);
2077 /* Try the LD_LIBRARY_PATH environment variable. */
2078 if (fd == -1 && env_path_list.dirs != (void *) -1)
2079 fd = open_path (name, namelen, preloaded, &env_path_list,
2080 &realname, &fb,
2081 loader ?: GL(dl_ns)[LM_ID_BASE]._ns_loaded,
2082 LA_SER_LIBPATH, &found_other_class);
2084 /* Look at the RUNPATH information for this binary. */
2085 if (fd == -1 && loader != NULL
2086 && cache_rpath (loader, &loader->l_runpath_dirs,
2087 DT_RUNPATH, "RUNPATH"))
2088 fd = open_path (name, namelen, preloaded,
2089 &loader->l_runpath_dirs, &realname, &fb, loader,
2090 LA_SER_RUNPATH, &found_other_class);
2092 if (fd == -1
2093 && (__builtin_expect (! preloaded, 1)
2094 || ! INTUSE(__libc_enable_secure)))
2096 /* Check the list of libraries in the file /etc/ld.so.cache,
2097 for compatibility with Linux's ldconfig program. */
2098 const char *cached = _dl_load_cache_lookup (name);
2100 if (cached != NULL)
2102 #ifdef SHARED
2103 // XXX Correct to unconditionally default to namespace 0?
2104 l = loader ?: GL(dl_ns)[LM_ID_BASE]._ns_loaded;
2105 #else
2106 l = loader;
2107 #endif
2109 /* If the loader has the DF_1_NODEFLIB flag set we must not
2110 use a cache entry from any of these directories. */
2111 if (
2112 #ifndef SHARED
2113 /* 'l' is always != NULL for dynamically linked objects. */
2114 l != NULL &&
2115 #endif
2116 __builtin_expect (l->l_flags_1 & DF_1_NODEFLIB, 0))
2118 const char *dirp = system_dirs;
2119 unsigned int cnt = 0;
2123 if (memcmp (cached, dirp, system_dirs_len[cnt]) == 0)
2125 /* The prefix matches. Don't use the entry. */
2126 cached = NULL;
2127 break;
2130 dirp += system_dirs_len[cnt] + 1;
2131 ++cnt;
2133 while (cnt < nsystem_dirs_len);
2136 if (cached != NULL)
2138 fd = open_verify (cached,
2139 &fb, loader ?: GL(dl_ns)[nsid]._ns_loaded,
2140 LA_SER_CONFIG, &found_other_class, false);
2141 if (__builtin_expect (fd != -1, 1))
2143 realname = local_strdup (cached);
2144 if (realname == NULL)
2146 __close (fd);
2147 fd = -1;
2154 /* Finally, try the default path. */
2155 if (fd == -1
2156 && ((l = loader ?: GL(dl_ns)[nsid]._ns_loaded) == NULL
2157 || __builtin_expect (!(l->l_flags_1 & DF_1_NODEFLIB), 1))
2158 && rtld_search_dirs.dirs != (void *) -1)
2159 fd = open_path (name, namelen, preloaded, &rtld_search_dirs,
2160 &realname, &fb, l, LA_SER_DEFAULT, &found_other_class);
2162 /* Add another newline when we are tracing the library loading. */
2163 if (__builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_LIBS, 0))
2164 _dl_debug_printf ("\n");
2166 else
2168 /* The path may contain dynamic string tokens. */
2169 realname = (loader
2170 ? expand_dynamic_string_token (loader, name)
2171 : local_strdup (name));
2172 if (realname == NULL)
2173 fd = -1;
2174 else
2176 fd = open_verify (realname, &fb,
2177 loader ?: GL(dl_ns)[nsid]._ns_loaded, 0,
2178 &found_other_class, true);
2179 if (__builtin_expect (fd, 0) == -1)
2180 free (realname);
2184 #ifdef SHARED
2185 no_file:
2186 #endif
2187 /* In case the LOADER information has only been provided to get to
2188 the appropriate RUNPATH/RPATH information we do not need it
2189 anymore. */
2190 if (mode & __RTLD_CALLMAP)
2191 loader = NULL;
2193 if (__builtin_expect (fd, 0) == -1)
2195 if (trace_mode
2196 && __builtin_expect (GLRO(dl_debug_mask) & DL_DEBUG_PRELINK, 0) == 0)
2198 /* We haven't found an appropriate library. But since we
2199 are only interested in the list of libraries this isn't
2200 so severe. Fake an entry with all the information we
2201 have. */
2202 static const Elf_Symndx dummy_bucket = STN_UNDEF;
2204 /* Enter the new object in the list of loaded objects. */
2205 if ((name_copy = local_strdup (name)) == NULL
2206 || (l = _dl_new_object (name_copy, name, type, loader,
2207 mode, nsid)) == NULL)
2209 free (name_copy);
2210 _dl_signal_error (ENOMEM, name, NULL,
2211 N_("cannot create shared object descriptor"));
2213 /* Signal that this is a faked entry. */
2214 l->l_faked = 1;
2215 /* Since the descriptor is initialized with zero we do not
2216 have do this here.
2217 l->l_reserved = 0; */
2218 l->l_buckets = &dummy_bucket;
2219 l->l_nbuckets = 1;
2220 l->l_relocated = 1;
2222 return l;
2224 else if (found_other_class)
2225 _dl_signal_error (0, name, NULL,
2226 ELFW(CLASS) == ELFCLASS32
2227 ? N_("wrong ELF class: ELFCLASS64")
2228 : N_("wrong ELF class: ELFCLASS32"));
2229 else
2230 _dl_signal_error (errno, name, NULL,
2231 N_("cannot open shared object file"));
2234 void *stack_end = __libc_stack_end;
2235 return _dl_map_object_from_fd (name, fd, &fb, realname, loader, type, mode,
2236 &stack_end, nsid);
2240 void
2241 internal_function
2242 _dl_rtld_di_serinfo (struct link_map *loader, Dl_serinfo *si, bool counting)
2244 if (counting)
2246 si->dls_cnt = 0;
2247 si->dls_size = 0;
2250 unsigned int idx = 0;
2251 char *allocptr = (char *) &si->dls_serpath[si->dls_cnt];
2252 void add_path (const struct r_search_path_struct *sps, unsigned int flags)
2253 # define add_path(sps, flags) add_path(sps, 0) /* XXX */
2255 if (sps->dirs != (void *) -1)
2257 struct r_search_path_elem **dirs = sps->dirs;
2260 const struct r_search_path_elem *const r = *dirs++;
2261 if (counting)
2263 si->dls_cnt++;
2264 si->dls_size += MAX (2, r->dirnamelen);
2266 else
2268 Dl_serpath *const sp = &si->dls_serpath[idx++];
2269 sp->dls_name = allocptr;
2270 if (r->dirnamelen < 2)
2271 *allocptr++ = r->dirnamelen ? '/' : '.';
2272 else
2273 allocptr = __mempcpy (allocptr,
2274 r->dirname, r->dirnamelen - 1);
2275 *allocptr++ = '\0';
2276 sp->dls_flags = flags;
2279 while (*dirs != NULL);
2283 /* When the object has the RUNPATH information we don't use any RPATHs. */
2284 if (loader->l_info[DT_RUNPATH] == NULL)
2286 /* First try the DT_RPATH of the dependent object that caused NAME
2287 to be loaded. Then that object's dependent, and on up. */
2289 struct link_map *l = loader;
2292 if (cache_rpath (l, &l->l_rpath_dirs, DT_RPATH, "RPATH"))
2293 add_path (&l->l_rpath_dirs, XXX_RPATH);
2294 l = l->l_loader;
2296 while (l != NULL);
2298 /* If dynamically linked, try the DT_RPATH of the executable itself. */
2299 if (loader->l_ns == LM_ID_BASE)
2301 l = GL(dl_ns)[LM_ID_BASE]._ns_loaded;
2302 if (l != NULL && l->l_type != lt_loaded && l != loader)
2303 if (cache_rpath (l, &l->l_rpath_dirs, DT_RPATH, "RPATH"))
2304 add_path (&l->l_rpath_dirs, XXX_RPATH);
2308 /* Try the LD_LIBRARY_PATH environment variable. */
2309 add_path (&env_path_list, XXX_ENV);
2311 /* Look at the RUNPATH information for this binary. */
2312 if (cache_rpath (loader, &loader->l_runpath_dirs, DT_RUNPATH, "RUNPATH"))
2313 add_path (&loader->l_runpath_dirs, XXX_RUNPATH);
2315 /* XXX
2316 Here is where ld.so.cache gets checked, but we don't have
2317 a way to indicate that in the results for Dl_serinfo. */
2319 /* Finally, try the default path. */
2320 if (!(loader->l_flags_1 & DF_1_NODEFLIB))
2321 add_path (&rtld_search_dirs, XXX_default);
2323 if (counting)
2324 /* Count the struct size before the string area, which we didn't
2325 know before we completed dls_cnt. */
2326 si->dls_size += (char *) &si->dls_serpath[si->dls_cnt] - (char *) si;