1/*
2 * Copyright (C) 2008, 2009 The Android Open Source Project
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 *  * Redistributions of source code must retain the above copyright
9 *    notice, this list of conditions and the following disclaimer.
10 *  * Redistributions in binary form must reproduce the above copyright
11 *    notice, this list of conditions and the following disclaimer in
12 *    the documentation and/or other materials provided with the
13 *    distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
16 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
17 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
18 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
19 * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
21 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
22 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
23 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
24 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
25 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * SUCH DAMAGE.
27 */
28
29#include <linux/auxvec.h>
30
31#include <stdio.h>
32#include <stdlib.h>
33#include <string.h>
34#include <unistd.h>
35#include <fcntl.h>
36#include <errno.h>
37#include <dlfcn.h>
38#include <sys/stat.h>
39
40#include <pthread.h>
41
42#include <sys/mman.h>
43
44#include <sys/atomics.h>
45
46/* special private C library header - see Android.mk */
47#include <bionic_tls.h>
48
49#include "linker.h"
50#include "linker_debug.h"
51#include "linker_environ.h"
52#include "linker_format.h"
53
54#define ALLOW_SYMBOLS_FROM_MAIN 1
55#define SO_MAX 128
56
57/* Assume average path length of 64 and max 8 paths */
58#define LDPATH_BUFSIZE 512
59#define LDPATH_MAX 8
60
61#define LDPRELOAD_BUFSIZE 512
62#define LDPRELOAD_MAX 8
63
64/* >>> IMPORTANT NOTE - READ ME BEFORE MODIFYING <<<
65 *
66 * Do NOT use malloc() and friends or pthread_*() code here.
67 * Don't use printf() either; it's caused mysterious memory
68 * corruption in the past.
69 * The linker runs before we bring up libc and it's easiest
70 * to make sure it does not depend on any complex libc features
71 *
72 * open issues / todo:
73 *
74 * - are we doing everything we should for ARM_COPY relocations?
75 * - cleaner error reporting
76 * - after linking, set as much stuff as possible to READONLY
77 *   and NOEXEC
78 * - linker hardcodes PAGE_SIZE and PAGE_MASK because the kernel
79 *   headers provide versions that are negative...
80 * - allocate space for soinfo structs dynamically instead of
81 *   having a hard limit (64)
82*/
83
84
85static int link_image(soinfo *si, unsigned wr_offset);
86
87static int socount = 0;
88static soinfo sopool[SO_MAX];
89static soinfo *freelist = NULL;
90static soinfo *solist = &libdl_info;
91static soinfo *sonext = &libdl_info;
92#if ALLOW_SYMBOLS_FROM_MAIN
93static soinfo *somain; /* main process, always the one after libdl_info */
94#endif
95
96
97static inline int validate_soinfo(soinfo *si)
98{
99    return (si >= sopool && si < sopool + SO_MAX) ||
100        si == &libdl_info;
101}
102
103static char ldpaths_buf[LDPATH_BUFSIZE];
104static const char *ldpaths[LDPATH_MAX + 1];
105
106static char ldpreloads_buf[LDPRELOAD_BUFSIZE];
107static const char *ldpreload_names[LDPRELOAD_MAX + 1];
108
109static soinfo *preloads[LDPRELOAD_MAX + 1];
110
111#if LINKER_DEBUG
112int debug_verbosity;
113#endif
114
115static int pid;
116
117/* This boolean is set if the program being loaded is setuid */
118static int program_is_setuid;
119
120#if STATS
121struct _link_stats linker_stats;
122#endif
123
124#if COUNT_PAGES
125unsigned bitmask[4096];
126#endif
127
128#ifndef PT_ARM_EXIDX
129#define PT_ARM_EXIDX    0x70000001      /* .ARM.exidx segment */
130#endif
131
132#define HOODLUM(name, ret, ...)                                               \
133    ret name __VA_ARGS__                                                      \
134    {                                                                         \
135        char errstr[] = "ERROR: " #name " called from the dynamic linker!\n"; \
136        write(2, errstr, sizeof(errstr));                                     \
137        abort();                                                              \
138    }
139HOODLUM(malloc, void *, (size_t size));
140HOODLUM(free, void, (void *ptr));
141HOODLUM(realloc, void *, (void *ptr, size_t size));
142HOODLUM(calloc, void *, (size_t cnt, size_t size));
143
144static char tmp_err_buf[768];
145static char __linker_dl_err_buf[768];
146#define DL_ERR(fmt, x...)                                                     \
147    do {                                                                      \
148        format_buffer(__linker_dl_err_buf, sizeof(__linker_dl_err_buf),            \
149                 "%s[%d]: " fmt, __func__, __LINE__, ##x);                    \
150        ERROR(fmt "\n", ##x);                                                      \
151    } while(0)
152
153const char *linker_get_error(void)
154{
155    return (const char *)&__linker_dl_err_buf[0];
156}
157
158/*
159 * This function is an empty stub where GDB locates a breakpoint to get notified
160 * about linker activity.
161 */
162extern void __attribute__((noinline)) __attribute__((visibility("default"))) rtld_db_dlactivity(void);
163
164static struct r_debug _r_debug = {1, NULL, &rtld_db_dlactivity,
165                                  RT_CONSISTENT, 0};
166static struct link_map *r_debug_tail = 0;
167
168static pthread_mutex_t _r_debug_lock = PTHREAD_MUTEX_INITIALIZER;
169
170static void insert_soinfo_into_debug_map(soinfo * info)
171{
172    struct link_map * map;
173
174    /* Copy the necessary fields into the debug structure.
175     */
176    map = &(info->linkmap);
177    map->l_addr = info->base;
178    map->l_name = (char*) info->name;
179    map->l_ld = (uintptr_t)info->dynamic;
180
181    /* Stick the new library at the end of the list.
182     * gdb tends to care more about libc than it does
183     * about leaf libraries, and ordering it this way
184     * reduces the back-and-forth over the wire.
185     */
186    if (r_debug_tail) {
187        r_debug_tail->l_next = map;
188        map->l_prev = r_debug_tail;
189        map->l_next = 0;
190    } else {
191        _r_debug.r_map = map;
192        map->l_prev = 0;
193        map->l_next = 0;
194    }
195    r_debug_tail = map;
196}
197
198static void remove_soinfo_from_debug_map(soinfo * info)
199{
200    struct link_map * map = &(info->linkmap);
201
202    if (r_debug_tail == map)
203        r_debug_tail = map->l_prev;
204
205    if (map->l_prev) map->l_prev->l_next = map->l_next;
206    if (map->l_next) map->l_next->l_prev = map->l_prev;
207}
208
209void notify_gdb_of_load(soinfo * info)
210{
211    if (info->flags & FLAG_EXE) {
212        // GDB already knows about the main executable
213        return;
214    }
215
216    pthread_mutex_lock(&_r_debug_lock);
217
218    _r_debug.r_state = RT_ADD;
219    rtld_db_dlactivity();
220
221    insert_soinfo_into_debug_map(info);
222
223    _r_debug.r_state = RT_CONSISTENT;
224    rtld_db_dlactivity();
225
226    pthread_mutex_unlock(&_r_debug_lock);
227}
228
229void notify_gdb_of_unload(soinfo * info)
230{
231    if (info->flags & FLAG_EXE) {
232        // GDB already knows about the main executable
233        return;
234    }
235
236    pthread_mutex_lock(&_r_debug_lock);
237
238    _r_debug.r_state = RT_DELETE;
239    rtld_db_dlactivity();
240
241    remove_soinfo_from_debug_map(info);
242
243    _r_debug.r_state = RT_CONSISTENT;
244    rtld_db_dlactivity();
245
246    pthread_mutex_unlock(&_r_debug_lock);
247}
248
249void notify_gdb_of_libraries()
250{
251    _r_debug.r_state = RT_ADD;
252    rtld_db_dlactivity();
253    _r_debug.r_state = RT_CONSISTENT;
254    rtld_db_dlactivity();
255}
256
257static soinfo *alloc_info(const char *name)
258{
259    soinfo *si;
260
261    if(strlen(name) >= SOINFO_NAME_LEN) {
262        DL_ERR("%5d library name %s too long", pid, name);
263        return NULL;
264    }
265
266    /* The freelist is populated when we call free_info(), which in turn is
267       done only by dlclose(), which is not likely to be used.
268    */
269    if (!freelist) {
270        if(socount == SO_MAX) {
271            DL_ERR("%5d too many libraries when loading %s", pid, name);
272            return NULL;
273        }
274        freelist = sopool + socount++;
275        freelist->next = NULL;
276    }
277
278    si = freelist;
279    freelist = freelist->next;
280
281    /* Make sure we get a clean block of soinfo */
282    memset(si, 0, sizeof(soinfo));
283    strlcpy((char*) si->name, name, sizeof(si->name));
284    sonext->next = si;
285    si->next = NULL;
286    si->refcount = 0;
287    sonext = si;
288
289    TRACE("%5d name %s: allocated soinfo @ %p\n", pid, name, si);
290    return si;
291}
292
293static void free_info(soinfo *si)
294{
295    soinfo *prev = NULL, *trav;
296
297    TRACE("%5d name %s: freeing soinfo @ %p\n", pid, si->name, si);
298
299    for(trav = solist; trav != NULL; trav = trav->next){
300        if (trav == si)
301            break;
302        prev = trav;
303    }
304    if (trav == NULL) {
305        /* si was not ni solist */
306        DL_ERR("%5d name %s is not in solist!", pid, si->name);
307        return;
308    }
309
310    /* prev will never be NULL, because the first entry in solist is
311       always the static libdl_info.
312    */
313    prev->next = si->next;
314    if (si == sonext) sonext = prev;
315    si->next = freelist;
316    freelist = si;
317}
318
319const char *addr_to_name(unsigned addr)
320{
321    soinfo *si;
322
323    for(si = solist; si != 0; si = si->next){
324        if((addr >= si->base) && (addr < (si->base + si->size))) {
325            return si->name;
326        }
327    }
328
329    return "";
330}
331
332/* For a given PC, find the .so that it belongs to.
333 * Returns the base address of the .ARM.exidx section
334 * for that .so, and the number of 8-byte entries
335 * in that section (via *pcount).
336 *
337 * Intended to be called by libc's __gnu_Unwind_Find_exidx().
338 *
339 * This function is exposed via dlfcn.c and libdl.so.
340 */
341#ifdef ANDROID_ARM_LINKER
342_Unwind_Ptr dl_unwind_find_exidx(_Unwind_Ptr pc, int *pcount)
343{
344    soinfo *si;
345    unsigned addr = (unsigned)pc;
346
347    for (si = solist; si != 0; si = si->next){
348        if ((addr >= si->base) && (addr < (si->base + si->size))) {
349            *pcount = si->ARM_exidx_count;
350            return (_Unwind_Ptr)(si->base + (unsigned long)si->ARM_exidx);
351        }
352    }
353   *pcount = 0;
354    return NULL;
355}
356#elif defined(ANDROID_X86_LINKER)
357/* Here, we only have to provide a callback to iterate across all the
358 * loaded libraries. gcc_eh does the rest. */
359int
360dl_iterate_phdr(int (*cb)(struct dl_phdr_info *info, size_t size, void *data),
361                void *data)
362{
363    soinfo *si;
364    struct dl_phdr_info dl_info;
365    int rv = 0;
366
367    for (si = solist; si != NULL; si = si->next) {
368        dl_info.dlpi_addr = si->linkmap.l_addr;
369        dl_info.dlpi_name = si->linkmap.l_name;
370        dl_info.dlpi_phdr = si->phdr;
371        dl_info.dlpi_phnum = si->phnum;
372        rv = cb(&dl_info, sizeof (struct dl_phdr_info), data);
373        if (rv != 0)
374            break;
375    }
376    return rv;
377}
378#endif
379
380static Elf32_Sym *_elf_lookup(soinfo *si, unsigned hash, const char *name)
381{
382    Elf32_Sym *s;
383    Elf32_Sym *symtab = si->symtab;
384    const char *strtab = si->strtab;
385    unsigned n;
386
387    TRACE_TYPE(LOOKUP, "%5d SEARCH %s in %s@0x%08x %08x %d\n", pid,
388               name, si->name, si->base, hash, hash % si->nbucket);
389    n = hash % si->nbucket;
390
391    for(n = si->bucket[hash % si->nbucket]; n != 0; n = si->chain[n]){
392        s = symtab + n;
393        if(strcmp(strtab + s->st_name, name)) continue;
394
395            /* only concern ourselves with global and weak symbol definitions */
396        switch(ELF32_ST_BIND(s->st_info)){
397        case STB_GLOBAL:
398        case STB_WEAK:
399                /* no section == undefined */
400            if(s->st_shndx == 0) continue;
401
402            TRACE_TYPE(LOOKUP, "%5d FOUND %s in %s (%08x) %d\n", pid,
403                       name, si->name, s->st_value, s->st_size);
404            return s;
405        }
406    }
407
408    return NULL;
409}
410
411static unsigned elfhash(const char *_name)
412{
413    const unsigned char *name = (const unsigned char *) _name;
414    unsigned h = 0, g;
415
416    while(*name) {
417        h = (h << 4) + *name++;
418        g = h & 0xf0000000;
419        h ^= g;
420        h ^= g >> 24;
421    }
422    return h;
423}
424
425static Elf32_Sym *
426_do_lookup(soinfo *si, const char *name, unsigned *base)
427{
428    unsigned elf_hash = elfhash(name);
429    Elf32_Sym *s;
430    unsigned *d;
431    soinfo *lsi = si;
432    int i;
433
434    /* Look for symbols in the local scope (the object who is
435     * searching). This happens with C++ templates on i386 for some
436     * reason.
437     *
438     * Notes on weak symbols:
439     * The ELF specs are ambigious about treatment of weak definitions in
440     * dynamic linking.  Some systems return the first definition found
441     * and some the first non-weak definition.   This is system dependent.
442     * Here we return the first definition found for simplicity.  */
443
444    s = _elf_lookup(si, elf_hash, name);
445    if(s != NULL)
446        goto done;
447
448    /* Next, look for it in the preloads list */
449    for(i = 0; preloads[i] != NULL; i++) {
450        lsi = preloads[i];
451        s = _elf_lookup(lsi, elf_hash, name);
452        if(s != NULL)
453            goto done;
454    }
455
456    for(d = si->dynamic; *d; d += 2) {
457        if(d[0] == DT_NEEDED){
458            lsi = (soinfo *)d[1];
459            if (!validate_soinfo(lsi)) {
460                DL_ERR("%5d bad DT_NEEDED pointer in %s",
461                       pid, si->name);
462                return NULL;
463            }
464
465            DEBUG("%5d %s: looking up %s in %s\n",
466                  pid, si->name, name, lsi->name);
467            s = _elf_lookup(lsi, elf_hash, name);
468            if ((s != NULL) && (s->st_shndx != SHN_UNDEF))
469                goto done;
470        }
471    }
472
473#if ALLOW_SYMBOLS_FROM_MAIN
474    /* If we are resolving relocations while dlopen()ing a library, it's OK for
475     * the library to resolve a symbol that's defined in the executable itself,
476     * although this is rare and is generally a bad idea.
477     */
478    if (somain) {
479        lsi = somain;
480        DEBUG("%5d %s: looking up %s in executable %s\n",
481              pid, si->name, name, lsi->name);
482        s = _elf_lookup(lsi, elf_hash, name);
483    }
484#endif
485
486done:
487    if(s != NULL) {
488        TRACE_TYPE(LOOKUP, "%5d si %s sym %s s->st_value = 0x%08x, "
489                   "found in %s, base = 0x%08x\n",
490                   pid, si->name, name, s->st_value, lsi->name, lsi->base);
491        *base = lsi->base;
492        return s;
493    }
494
495    return NULL;
496}
497
498/* This is used by dl_sym().  It performs symbol lookup only within the
499   specified soinfo object and not in any of its dependencies.
500 */
501Elf32_Sym *lookup_in_library(soinfo *si, const char *name)
502{
503    return _elf_lookup(si, elfhash(name), name);
504}
505
506/* This is used by dl_sym().  It performs a global symbol lookup.
507 */
508Elf32_Sym *lookup(const char *name, soinfo **found, soinfo *start)
509{
510    unsigned elf_hash = elfhash(name);
511    Elf32_Sym *s = NULL;
512    soinfo *si;
513
514    if(start == NULL) {
515        start = solist;
516    }
517
518    for(si = start; (s == NULL) && (si != NULL); si = si->next)
519    {
520        if(si->flags & FLAG_ERROR)
521            continue;
522        s = _elf_lookup(si, elf_hash, name);
523        if (s != NULL) {
524            *found = si;
525            break;
526        }
527    }
528
529    if(s != NULL) {
530        TRACE_TYPE(LOOKUP, "%5d %s s->st_value = 0x%08x, "
531                   "si->base = 0x%08x\n", pid, name, s->st_value, si->base);
532        return s;
533    }
534
535    return NULL;
536}
537
538soinfo *find_containing_library(const void *addr)
539{
540    soinfo *si;
541
542    for(si = solist; si != NULL; si = si->next)
543    {
544        if((unsigned)addr >= si->base && (unsigned)addr - si->base < si->size) {
545            return si;
546        }
547    }
548
549    return NULL;
550}
551
552Elf32_Sym *find_containing_symbol(const void *addr, soinfo *si)
553{
554    unsigned int i;
555    unsigned soaddr = (unsigned)addr - si->base;
556
557    /* Search the library's symbol table for any defined symbol which
558     * contains this address */
559    for(i=0; i<si->nchain; i++) {
560        Elf32_Sym *sym = &si->symtab[i];
561
562        if(sym->st_shndx != SHN_UNDEF &&
563           soaddr >= sym->st_value &&
564           soaddr < sym->st_value + sym->st_size) {
565            return sym;
566        }
567    }
568
569    return NULL;
570}
571
572#if 0
573static void dump(soinfo *si)
574{
575    Elf32_Sym *s = si->symtab;
576    unsigned n;
577
578    for(n = 0; n < si->nchain; n++) {
579        TRACE("%5d %04d> %08x: %02x %04x %08x %08x %s\n", pid, n, s,
580               s->st_info, s->st_shndx, s->st_value, s->st_size,
581               si->strtab + s->st_name);
582        s++;
583    }
584}
585#endif
586
587static const char *sopaths[] = {
588    "/vendor/lib",
589    "/system/lib",
590    0
591};
592
593static int _open_lib(const char *name)
594{
595    int fd;
596    struct stat filestat;
597
598    if ((stat(name, &filestat) >= 0) && S_ISREG(filestat.st_mode)) {
599        if ((fd = open(name, O_RDONLY)) >= 0)
600            return fd;
601    }
602
603    return -1;
604}
605
606static int open_library(const char *name)
607{
608    int fd;
609    char buf[512];
610    const char **path;
611    int n;
612
613    TRACE("[ %5d opening %s ]\n", pid, name);
614
615    if(name == 0) return -1;
616    if(strlen(name) > 256) return -1;
617
618    if ((name[0] == '/') && ((fd = _open_lib(name)) >= 0))
619        return fd;
620
621    for (path = ldpaths; *path; path++) {
622        n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
623        if (n < 0 || n >= (int)sizeof(buf)) {
624            WARN("Ignoring very long library path: %s/%s\n", *path, name);
625            continue;
626        }
627        if ((fd = _open_lib(buf)) >= 0)
628            return fd;
629    }
630    for (path = sopaths; *path; path++) {
631        n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
632        if (n < 0 || n >= (int)sizeof(buf)) {
633            WARN("Ignoring very long library path: %s/%s\n", *path, name);
634            continue;
635        }
636        if ((fd = _open_lib(buf)) >= 0)
637            return fd;
638    }
639
640    return -1;
641}
642
643/* temporary space for holding the first page of the shared lib
644 * which contains the elf header (with the pht). */
645static unsigned char __header[PAGE_SIZE];
646
647typedef struct {
648    long mmap_addr;
649    char tag[4]; /* 'P', 'R', 'E', ' ' */
650} prelink_info_t;
651
652/* Returns the requested base address if the library is prelinked,
653 * and 0 otherwise.  */
654static unsigned long
655is_prelinked(int fd, const char *name)
656{
657    off_t sz;
658    prelink_info_t info;
659
660    sz = lseek(fd, -sizeof(prelink_info_t), SEEK_END);
661    if (sz < 0) {
662        DL_ERR("lseek() failed!");
663        return 0;
664    }
665
666    if (read(fd, &info, sizeof(info)) != sizeof(info)) {
667        WARN("Could not read prelink_info_t structure for `%s`\n", name);
668        return 0;
669    }
670
671    if (strncmp(info.tag, "PRE ", 4)) {
672        WARN("`%s` is not a prelinked library\n", name);
673        return 0;
674    }
675
676    return (unsigned long)info.mmap_addr;
677}
678
679/* verify_elf_object
680 *      Verifies if the object @ base is a valid ELF object
681 *
682 * Args:
683 *
684 * Returns:
685 *       0 on success
686 *      -1 if no valid ELF object is found @ base.
687 */
688static int
689verify_elf_object(void *base, const char *name)
690{
691    Elf32_Ehdr *hdr = (Elf32_Ehdr *) base;
692
693    if (hdr->e_ident[EI_MAG0] != ELFMAG0) return -1;
694    if (hdr->e_ident[EI_MAG1] != ELFMAG1) return -1;
695    if (hdr->e_ident[EI_MAG2] != ELFMAG2) return -1;
696    if (hdr->e_ident[EI_MAG3] != ELFMAG3) return -1;
697
698    /* TODO: Should we verify anything else in the header? */
699#ifdef ANDROID_ARM_LINKER
700    if (hdr->e_machine != EM_ARM) return -1;
701#elif defined(ANDROID_X86_LINKER)
702    if (hdr->e_machine != EM_386) return -1;
703#endif
704    return 0;
705}
706
707
708/* get_lib_extents
709 *      Retrieves the base (*base) address where the ELF object should be
710 *      mapped and its overall memory size (*total_sz).
711 *
712 * Args:
713 *      fd: Opened file descriptor for the library
714 *      name: The name of the library
715 *      _hdr: Pointer to the header page of the library
716 *      total_sz: Total size of the memory that should be allocated for
717 *                this library
718 *
719 * Returns:
720 *      -1 if there was an error while trying to get the lib extents.
721 *         The possible reasons are:
722 *             - Could not determine if the library was prelinked.
723 *             - The library provided is not a valid ELF object
724 *       0 if the library did not request a specific base offset (normal
725 *         for non-prelinked libs)
726 *     > 0 if the library requests a specific address to be mapped to.
727 *         This indicates a pre-linked library.
728 */
729static unsigned
730get_lib_extents(int fd, const char *name, void *__hdr, unsigned *total_sz)
731{
732    unsigned req_base;
733    unsigned min_vaddr = 0xffffffff;
734    unsigned max_vaddr = 0;
735    unsigned char *_hdr = (unsigned char *)__hdr;
736    Elf32_Ehdr *ehdr = (Elf32_Ehdr *)_hdr;
737    Elf32_Phdr *phdr;
738    int cnt;
739
740    TRACE("[ %5d Computing extents for '%s'. ]\n", pid, name);
741    if (verify_elf_object(_hdr, name) < 0) {
742        DL_ERR("%5d - %s is not a valid ELF object", pid, name);
743        return (unsigned)-1;
744    }
745
746    req_base = (unsigned) is_prelinked(fd, name);
747    if (req_base == (unsigned)-1)
748        return -1;
749    else if (req_base != 0) {
750        TRACE("[ %5d - Prelinked library '%s' requesting base @ 0x%08x ]\n",
751              pid, name, req_base);
752    } else {
753        TRACE("[ %5d - Non-prelinked library '%s' found. ]\n", pid, name);
754    }
755
756    phdr = (Elf32_Phdr *)(_hdr + ehdr->e_phoff);
757
758    /* find the min/max p_vaddrs from all the PT_LOAD segments so we can
759     * get the range. */
760    for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
761        if (phdr->p_type == PT_LOAD) {
762            if ((phdr->p_vaddr + phdr->p_memsz) > max_vaddr)
763                max_vaddr = phdr->p_vaddr + phdr->p_memsz;
764            if (phdr->p_vaddr < min_vaddr)
765                min_vaddr = phdr->p_vaddr;
766        }
767    }
768
769    if ((min_vaddr == 0xffffffff) && (max_vaddr == 0)) {
770        DL_ERR("%5d - No loadable segments found in %s.", pid, name);
771        return (unsigned)-1;
772    }
773
774    /* truncate min_vaddr down to page boundary */
775    min_vaddr &= ~PAGE_MASK;
776
777    /* round max_vaddr up to the next page */
778    max_vaddr = (max_vaddr + PAGE_SIZE - 1) & ~PAGE_MASK;
779
780    *total_sz = (max_vaddr - min_vaddr);
781    return (unsigned)req_base;
782}
783
784/* reserve_mem_region
785 *
786 *     This function reserves a chunk of memory to be used for mapping in
787 *     a prelinked shared library. We reserve the entire memory region here, and
788 *     then the rest of the linker will relocate the individual loadable
789 *     segments into the correct locations within this memory range.
790 *
791 * Args:
792 *     si->base: The requested base of the allocation.
793 *     si->size: The size of the allocation.
794 *
795 * Returns:
796 *     -1 on failure, and 0 on success.  On success, si->base will contain
797 *     the virtual address at which the library will be mapped.
798 */
799
800static int reserve_mem_region(soinfo *si)
801{
802    void *base = mmap((void *)si->base, si->size, PROT_NONE,
803                      MAP_FIXED | MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
804    if (base == MAP_FAILED) {
805        DL_ERR("%5d can NOT map (%sprelinked) library '%s' at 0x%08x "
806              "as requested, will try general pool: %d (%s)",
807              pid, (si->base ? "" : "non-"), si->name, si->base,
808              errno, strerror(errno));
809        return -1;
810    } else if (base != (void *)si->base) {
811        DL_ERR("OOPS: %5d %sprelinked library '%s' mapped at 0x%08x, "
812              "not at 0x%08x", pid, (si->base ? "" : "non-"),
813              si->name, (unsigned)base, si->base);
814        munmap(base, si->size);
815        return -1;
816    }
817    return 0;
818}
819
820static int alloc_mem_region(soinfo *si)
821{
822    if (si->base) {
823        /* Attempt to mmap a prelinked library. */
824        return reserve_mem_region(si);
825    }
826
827    /* This is not a prelinked library, so we use the kernel's default
828       allocator.
829    */
830
831    void *base = mmap(NULL, si->size, PROT_NONE,
832                      MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
833    if (base == MAP_FAILED) {
834        DL_ERR("%5d mmap of library '%s' failed: %d (%s)\n",
835              pid, si->name,
836              errno, strerror(errno));
837        goto err;
838    }
839    si->base = (unsigned) base;
840    PRINT("%5d mapped library '%s' to %08x via kernel allocator.\n",
841          pid, si->name, si->base);
842    return 0;
843
844err:
845    DL_ERR("OOPS: %5d cannot map library '%s'. no vspace available.",
846          pid, si->name);
847    return -1;
848}
849
850#define MAYBE_MAP_FLAG(x,from,to)    (((x) & (from)) ? (to) : 0)
851#define PFLAGS_TO_PROT(x)            (MAYBE_MAP_FLAG((x), PF_X, PROT_EXEC) | \
852                                      MAYBE_MAP_FLAG((x), PF_R, PROT_READ) | \
853                                      MAYBE_MAP_FLAG((x), PF_W, PROT_WRITE))
854/* load_segments
855 *
856 *     This function loads all the loadable (PT_LOAD) segments into memory
857 *     at their appropriate memory offsets off the base address.
858 *
859 * Args:
860 *     fd: Open file descriptor to the library to load.
861 *     header: Pointer to a header page that contains the ELF header.
862 *             This is needed since we haven't mapped in the real file yet.
863 *     si: ptr to soinfo struct describing the shared object.
864 *
865 * Returns:
866 *     0 on success, -1 on failure.
867 */
868static int
869load_segments(int fd, void *header, soinfo *si)
870{
871    Elf32_Ehdr *ehdr = (Elf32_Ehdr *)header;
872    Elf32_Phdr *phdr = (Elf32_Phdr *)((unsigned char *)header + ehdr->e_phoff);
873    Elf32_Addr base = (Elf32_Addr) si->base;
874    int cnt;
875    unsigned len;
876    Elf32_Addr tmp;
877    unsigned char *pbase;
878    unsigned char *extra_base;
879    unsigned extra_len;
880    unsigned total_sz = 0;
881
882    si->wrprotect_start = 0xffffffff;
883    si->wrprotect_end = 0;
884
885    TRACE("[ %5d - Begin loading segments for '%s' @ 0x%08x ]\n",
886          pid, si->name, (unsigned)si->base);
887    /* Now go through all the PT_LOAD segments and map them into memory
888     * at the appropriate locations. */
889    for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
890        if (phdr->p_type == PT_LOAD) {
891            DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
892            /* we want to map in the segment on a page boundary */
893            tmp = base + (phdr->p_vaddr & (~PAGE_MASK));
894            /* add the # of bytes we masked off above to the total length. */
895            len = phdr->p_filesz + (phdr->p_vaddr & PAGE_MASK);
896
897            TRACE("[ %d - Trying to load segment from '%s' @ 0x%08x "
898                  "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x ]\n", pid, si->name,
899                  (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
900            pbase = mmap((void *)tmp, len, PFLAGS_TO_PROT(phdr->p_flags),
901                         MAP_PRIVATE | MAP_FIXED, fd,
902                         phdr->p_offset & (~PAGE_MASK));
903            if (pbase == MAP_FAILED) {
904                DL_ERR("%d failed to map segment from '%s' @ 0x%08x (0x%08x). "
905                      "p_vaddr=0x%08x p_offset=0x%08x", pid, si->name,
906                      (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
907                goto fail;
908            }
909
910            /* If 'len' didn't end on page boundary, and it's a writable
911             * segment, zero-fill the rest. */
912            if ((len & PAGE_MASK) && (phdr->p_flags & PF_W))
913                memset((void *)(pbase + len), 0, PAGE_SIZE - (len & PAGE_MASK));
914
915            /* Check to see if we need to extend the map for this segment to
916             * cover the diff between filesz and memsz (i.e. for bss).
917             *
918             *  base           _+---------------------+  page boundary
919             *                  .                     .
920             *                  |                     |
921             *                  .                     .
922             *  pbase          _+---------------------+  page boundary
923             *                  |                     |
924             *                  .                     .
925             *  base + p_vaddr _|                     |
926             *                  . \          \        .
927             *                  . | filesz   |        .
928             *  pbase + len    _| /          |        |
929             *     <0 pad>      .            .        .
930             *  extra_base     _+------------|--------+  page boundary
931             *               /  .            .        .
932             *               |  .            .        .
933             *               |  +------------|--------+  page boundary
934             *  extra_len->  |  |            |        |
935             *               |  .            | memsz  .
936             *               |  .            |        .
937             *               \ _|            /        |
938             *                  .                     .
939             *                  |                     |
940             *                 _+---------------------+  page boundary
941             */
942            tmp = (Elf32_Addr)(((unsigned)pbase + len + PAGE_SIZE - 1) &
943                                    (~PAGE_MASK));
944            if (tmp < (base + phdr->p_vaddr + phdr->p_memsz)) {
945                extra_len = base + phdr->p_vaddr + phdr->p_memsz - tmp;
946                TRACE("[ %5d - Need to extend segment from '%s' @ 0x%08x "
947                      "(0x%08x) ]\n", pid, si->name, (unsigned)tmp, extra_len);
948                /* map in the extra page(s) as anonymous into the range.
949                 * This is probably not necessary as we already mapped in
950                 * the entire region previously, but we just want to be
951                 * sure. This will also set the right flags on the region
952                 * (though we can probably accomplish the same thing with
953                 * mprotect).
954                 */
955                extra_base = mmap((void *)tmp, extra_len,
956                                  PFLAGS_TO_PROT(phdr->p_flags),
957                                  MAP_PRIVATE | MAP_FIXED | MAP_ANONYMOUS,
958                                  -1, 0);
959                if (extra_base == MAP_FAILED) {
960                    DL_ERR("[ %5d - failed to extend segment from '%s' @ 0x%08x"
961                           " (0x%08x) ]", pid, si->name, (unsigned)tmp,
962                          extra_len);
963                    goto fail;
964                }
965                /* TODO: Check if we need to memset-0 this region.
966                 * Anonymous mappings are zero-filled copy-on-writes, so we
967                 * shouldn't need to. */
968                TRACE("[ %5d - Segment from '%s' extended @ 0x%08x "
969                      "(0x%08x)\n", pid, si->name, (unsigned)extra_base,
970                      extra_len);
971            }
972            /* set the len here to show the full extent of the segment we
973             * just loaded, mostly for debugging */
974            len = (((unsigned)base + phdr->p_vaddr + phdr->p_memsz +
975                    PAGE_SIZE - 1) & (~PAGE_MASK)) - (unsigned)pbase;
976            TRACE("[ %5d - Successfully loaded segment from '%s' @ 0x%08x "
977                  "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x\n", pid, si->name,
978                  (unsigned)pbase, len, phdr->p_vaddr, phdr->p_offset);
979            total_sz += len;
980            /* Make the section writable just in case we'll have to write to
981             * it during relocation (i.e. text segment). However, we will
982             * remember what range of addresses should be write protected.
983             *
984             */
985            if (!(phdr->p_flags & PF_W)) {
986                if ((unsigned)pbase < si->wrprotect_start)
987                    si->wrprotect_start = (unsigned)pbase;
988                if (((unsigned)pbase + len) > si->wrprotect_end)
989                    si->wrprotect_end = (unsigned)pbase + len;
990                mprotect(pbase, len,
991                         PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
992            }
993        } else if (phdr->p_type == PT_DYNAMIC) {
994            DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
995            /* this segment contains the dynamic linking information */
996            si->dynamic = (unsigned *)(base + phdr->p_vaddr);
997        } else if (phdr->p_type == PT_GNU_RELRO) {
998            if ((phdr->p_vaddr >= si->size)
999                    || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
1000                    || ((base + phdr->p_vaddr + phdr->p_memsz) < base)) {
1001                DL_ERR("%d invalid GNU_RELRO in '%s' "
1002                       "p_vaddr=0x%08x p_memsz=0x%08x", pid, si->name,
1003                       phdr->p_vaddr, phdr->p_memsz);
1004                goto fail;
1005            }
1006            si->gnu_relro_start = (Elf32_Addr) (base + phdr->p_vaddr);
1007            si->gnu_relro_len = (unsigned) phdr->p_memsz;
1008        } else {
1009#ifdef ANDROID_ARM_LINKER
1010            if (phdr->p_type == PT_ARM_EXIDX) {
1011                DEBUG_DUMP_PHDR(phdr, "PT_ARM_EXIDX", pid);
1012                /* exidx entries (used for stack unwinding) are 8 bytes each.
1013                 */
1014                si->ARM_exidx = (unsigned *)phdr->p_vaddr;
1015                si->ARM_exidx_count = phdr->p_memsz / 8;
1016            }
1017#endif
1018        }
1019
1020    }
1021
1022    /* Sanity check */
1023    if (total_sz > si->size) {
1024        DL_ERR("%5d - Total length (0x%08x) of mapped segments from '%s' is "
1025              "greater than what was allocated (0x%08x). THIS IS BAD!",
1026              pid, total_sz, si->name, si->size);
1027        goto fail;
1028    }
1029
1030    TRACE("[ %5d - Finish loading segments for '%s' @ 0x%08x. "
1031          "Total memory footprint: 0x%08x bytes ]\n", pid, si->name,
1032          (unsigned)si->base, si->size);
1033    return 0;
1034
1035fail:
1036    /* We can just blindly unmap the entire region even though some things
1037     * were mapped in originally with anonymous and others could have been
1038     * been mapped in from the file before we failed. The kernel will unmap
1039     * all the pages in the range, irrespective of how they got there.
1040     */
1041    munmap((void *)si->base, si->size);
1042    si->flags |= FLAG_ERROR;
1043    return -1;
1044}
1045
1046/* TODO: Implement this to take care of the fact that Android ARM
1047 * ELF objects shove everything into a single loadable segment that has the
1048 * write bit set. wr_offset is then used to set non-(data|bss) pages to be
1049 * non-writable.
1050 */
1051#if 0
1052static unsigned
1053get_wr_offset(int fd, const char *name, Elf32_Ehdr *ehdr)
1054{
1055    Elf32_Shdr *shdr_start;
1056    Elf32_Shdr *shdr;
1057    int shdr_sz = ehdr->e_shnum * sizeof(Elf32_Shdr);
1058    int cnt;
1059    unsigned wr_offset = 0xffffffff;
1060
1061    shdr_start = mmap(0, shdr_sz, PROT_READ, MAP_PRIVATE, fd,
1062                      ehdr->e_shoff & (~PAGE_MASK));
1063    if (shdr_start == MAP_FAILED) {
1064        WARN("%5d - Could not read section header info from '%s'. Will not "
1065             "not be able to determine write-protect offset.\n", pid, name);
1066        return (unsigned)-1;
1067    }
1068
1069    for(cnt = 0, shdr = shdr_start; cnt < ehdr->e_shnum; ++cnt, ++shdr) {
1070        if ((shdr->sh_type != SHT_NULL) && (shdr->sh_flags & SHF_WRITE) &&
1071            (shdr->sh_addr < wr_offset)) {
1072            wr_offset = shdr->sh_addr;
1073        }
1074    }
1075
1076    munmap(shdr_start, shdr_sz);
1077    return wr_offset;
1078}
1079#endif
1080
1081static soinfo *
1082load_library(const char *name)
1083{
1084    int fd = open_library(name);
1085    int cnt;
1086    unsigned ext_sz;
1087    unsigned req_base;
1088    const char *bname;
1089    soinfo *si = NULL;
1090    Elf32_Ehdr *hdr;
1091
1092    if(fd == -1) {
1093        DL_ERR("Library '%s' not found", name);
1094        return NULL;
1095    }
1096
1097    /* We have to read the ELF header to figure out what to do with this image
1098     */
1099    if (lseek(fd, 0, SEEK_SET) < 0) {
1100        DL_ERR("lseek() failed!");
1101        goto fail;
1102    }
1103
1104    if ((cnt = read(fd, &__header[0], PAGE_SIZE)) < 0) {
1105        DL_ERR("read() failed!");
1106        goto fail;
1107    }
1108
1109    /* Parse the ELF header and get the size of the memory footprint for
1110     * the library */
1111    req_base = get_lib_extents(fd, name, &__header[0], &ext_sz);
1112    if (req_base == (unsigned)-1)
1113        goto fail;
1114    TRACE("[ %5d - '%s' (%s) wants base=0x%08x sz=0x%08x ]\n", pid, name,
1115          (req_base ? "prelinked" : "not pre-linked"), req_base, ext_sz);
1116
1117    /* Now configure the soinfo struct where we'll store all of our data
1118     * for the ELF object. If the loading fails, we waste the entry, but
1119     * same thing would happen if we failed during linking. Configuring the
1120     * soinfo struct here is a lot more convenient.
1121     */
1122    bname = strrchr(name, '/');
1123    si = alloc_info(bname ? bname + 1 : name);
1124    if (si == NULL)
1125        goto fail;
1126
1127    /* Carve out a chunk of memory where we will map in the individual
1128     * segments */
1129    si->base = req_base;
1130    si->size = ext_sz;
1131    si->flags = 0;
1132    si->entry = 0;
1133    si->dynamic = (unsigned *)-1;
1134    if (alloc_mem_region(si) < 0)
1135        goto fail;
1136
1137    TRACE("[ %5d allocated memory for %s @ %p (0x%08x) ]\n",
1138          pid, name, (void *)si->base, (unsigned) ext_sz);
1139
1140    /* Now actually load the library's segments into right places in memory */
1141    if (load_segments(fd, &__header[0], si) < 0) {
1142        goto fail;
1143    }
1144
1145    /* this might not be right. Technically, we don't even need this info
1146     * once we go through 'load_segments'. */
1147    hdr = (Elf32_Ehdr *)si->base;
1148    si->phdr = (Elf32_Phdr *)((unsigned char *)si->base + hdr->e_phoff);
1149    si->phnum = hdr->e_phnum;
1150    /**/
1151
1152    close(fd);
1153    return si;
1154
1155fail:
1156    if (si) free_info(si);
1157    close(fd);
1158    return NULL;
1159}
1160
1161static soinfo *
1162init_library(soinfo *si)
1163{
1164    unsigned wr_offset = 0xffffffff;
1165
1166    /* At this point we know that whatever is loaded @ base is a valid ELF
1167     * shared library whose segments are properly mapped in. */
1168    TRACE("[ %5d init_library base=0x%08x sz=0x%08x name='%s') ]\n",
1169          pid, si->base, si->size, si->name);
1170
1171    if(link_image(si, wr_offset)) {
1172            /* We failed to link.  However, we can only restore libbase
1173            ** if no additional libraries have moved it since we updated it.
1174            */
1175        munmap((void *)si->base, si->size);
1176        return NULL;
1177    }
1178
1179    return si;
1180}
1181
1182soinfo *find_library(const char *name)
1183{
1184    soinfo *si;
1185    const char *bname;
1186
1187#if ALLOW_SYMBOLS_FROM_MAIN
1188    if (name == NULL)
1189        return somain;
1190#else
1191    if (name == NULL)
1192        return NULL;
1193#endif
1194
1195    bname = strrchr(name, '/');
1196    bname = bname ? bname + 1 : name;
1197
1198    for(si = solist; si != 0; si = si->next){
1199        if(!strcmp(bname, si->name)) {
1200            if(si->flags & FLAG_ERROR) {
1201                DL_ERR("%5d '%s' failed to load previously", pid, bname);
1202                return NULL;
1203            }
1204            if(si->flags & FLAG_LINKED) return si;
1205            DL_ERR("OOPS: %5d recursive link to '%s'", pid, si->name);
1206            return NULL;
1207        }
1208    }
1209
1210    TRACE("[ %5d '%s' has not been loaded yet.  Locating...]\n", pid, name);
1211    si = load_library(name);
1212    if(si == NULL)
1213        return NULL;
1214    return init_library(si);
1215}
1216
1217/* TODO:
1218 *   notify gdb of unload
1219 *   for non-prelinked libraries, find a way to decrement libbase
1220 */
1221static void call_destructors(soinfo *si);
1222unsigned unload_library(soinfo *si)
1223{
1224    unsigned *d;
1225    if (si->refcount == 1) {
1226        TRACE("%5d unloading '%s'\n", pid, si->name);
1227        call_destructors(si);
1228
1229        /*
1230         * Make sure that we undo the PT_GNU_RELRO protections we added
1231         * in link_image. This is needed to undo the DT_NEEDED hack below.
1232         */
1233        if ((si->gnu_relro_start != 0) && (si->gnu_relro_len != 0)) {
1234            Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
1235            unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
1236            if (mprotect((void *) start, len, PROT_READ | PROT_WRITE) < 0)
1237                DL_ERR("%5d %s: could not undo GNU_RELRO protections. "
1238                       "Expect a crash soon. errno=%d (%s)",
1239                       pid, si->name, errno, strerror(errno));
1240
1241        }
1242
1243        for(d = si->dynamic; *d; d += 2) {
1244            if(d[0] == DT_NEEDED){
1245                soinfo *lsi = (soinfo *)d[1];
1246
1247                // The next line will segfault if the we don't undo the
1248                // PT_GNU_RELRO protections (see comments above and in
1249                // link_image().
1250                d[1] = 0;
1251
1252                if (validate_soinfo(lsi)) {
1253                    TRACE("%5d %s needs to unload %s\n", pid,
1254                          si->name, lsi->name);
1255                    unload_library(lsi);
1256                }
1257                else
1258                    DL_ERR("%5d %s: could not unload dependent library",
1259                           pid, si->name);
1260            }
1261        }
1262
1263        munmap((char *)si->base, si->size);
1264        notify_gdb_of_unload(si);
1265        free_info(si);
1266        si->refcount = 0;
1267    }
1268    else {
1269        si->refcount--;
1270        PRINT("%5d not unloading '%s', decrementing refcount to %d\n",
1271              pid, si->name, si->refcount);
1272    }
1273    return si->refcount;
1274}
1275
1276/* TODO: don't use unsigned for addrs below. It works, but is not
1277 * ideal. They should probably be either uint32_t, Elf32_Addr, or unsigned
1278 * long.
1279 */
1280static int reloc_library(soinfo *si, Elf32_Rel *rel, unsigned count)
1281{
1282    Elf32_Sym *symtab = si->symtab;
1283    const char *strtab = si->strtab;
1284    Elf32_Sym *s;
1285    unsigned base;
1286    Elf32_Rel *start = rel;
1287    unsigned idx;
1288
1289    for (idx = 0; idx < count; ++idx) {
1290        unsigned type = ELF32_R_TYPE(rel->r_info);
1291        unsigned sym = ELF32_R_SYM(rel->r_info);
1292        unsigned reloc = (unsigned)(rel->r_offset + si->base);
1293        unsigned sym_addr = 0;
1294        char *sym_name = NULL;
1295
1296        DEBUG("%5d Processing '%s' relocation at index %d\n", pid,
1297              si->name, idx);
1298        if(sym != 0) {
1299            sym_name = (char *)(strtab + symtab[sym].st_name);
1300            s = _do_lookup(si, sym_name, &base);
1301            if(s == NULL) {
1302                /* We only allow an undefined symbol if this is a weak
1303                   reference..   */
1304                s = &symtab[sym];
1305                if (ELF32_ST_BIND(s->st_info) != STB_WEAK) {
1306                    DL_ERR("%5d cannot locate '%s'...\n", pid, sym_name);
1307                    return -1;
1308                }
1309
1310                /* IHI0044C AAELF 4.5.1.1:
1311
1312                   Libraries are not searched to resolve weak references.
1313                   It is not an error for a weak reference to remain
1314                   unsatisfied.
1315
1316                   During linking, the value of an undefined weak reference is:
1317                   - Zero if the relocation type is absolute
1318                   - The address of the place if the relocation is pc-relative
1319                   - The address of nominial base address if the relocation
1320                     type is base-relative.
1321                  */
1322
1323                switch (type) {
1324#if defined(ANDROID_ARM_LINKER)
1325                case R_ARM_JUMP_SLOT:
1326                case R_ARM_GLOB_DAT:
1327                case R_ARM_ABS32:
1328                case R_ARM_RELATIVE:    /* Don't care. */
1329                case R_ARM_NONE:        /* Don't care. */
1330#elif defined(ANDROID_X86_LINKER)
1331                case R_386_JUMP_SLOT:
1332                case R_386_GLOB_DAT:
1333                case R_386_32:
1334                case R_386_RELATIVE:    /* Dont' care. */
1335#endif /* ANDROID_*_LINKER */
1336                    /* sym_addr was initialized to be zero above or relocation
1337                       code below does not care about value of sym_addr.
1338                       No need to do anything.  */
1339                    break;
1340
1341#if defined(ANDROID_X86_LINKER)
1342                case R_386_PC32:
1343                    sym_addr = reloc;
1344                    break;
1345#endif /* ANDROID_X86_LINKER */
1346
1347#if defined(ANDROID_ARM_LINKER)
1348                case R_ARM_COPY:
1349                    /* Fall through.  Can't really copy if weak symbol is
1350                       not found in run-time.  */
1351#endif /* ANDROID_ARM_LINKER */
1352                default:
1353                    DL_ERR("%5d unknown weak reloc type %d @ %p (%d)\n",
1354                                 pid, type, rel, (int) (rel - start));
1355                    return -1;
1356                }
1357            } else {
1358                /* We got a definition.  */
1359#if 0
1360            if((base == 0) && (si->base != 0)){
1361                    /* linking from libraries to main image is bad */
1362                DL_ERR("%5d cannot locate '%s'...",
1363                       pid, strtab + symtab[sym].st_name);
1364                return -1;
1365            }
1366#endif
1367                sym_addr = (unsigned)(s->st_value + base);
1368	    }
1369            COUNT_RELOC(RELOC_SYMBOL);
1370        } else {
1371            s = NULL;
1372        }
1373
1374/* TODO: This is ugly. Split up the relocations by arch into
1375 * different files.
1376 */
1377        switch(type){
1378#if defined(ANDROID_ARM_LINKER)
1379        case R_ARM_JUMP_SLOT:
1380            COUNT_RELOC(RELOC_ABSOLUTE);
1381            MARK(rel->r_offset);
1382            TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
1383                       reloc, sym_addr, sym_name);
1384            *((unsigned*)reloc) = sym_addr;
1385            break;
1386        case R_ARM_GLOB_DAT:
1387            COUNT_RELOC(RELOC_ABSOLUTE);
1388            MARK(rel->r_offset);
1389            TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
1390                       reloc, sym_addr, sym_name);
1391            *((unsigned*)reloc) = sym_addr;
1392            break;
1393        case R_ARM_ABS32:
1394            COUNT_RELOC(RELOC_ABSOLUTE);
1395            MARK(rel->r_offset);
1396            TRACE_TYPE(RELO, "%5d RELO ABS %08x <- %08x %s\n", pid,
1397                       reloc, sym_addr, sym_name);
1398            *((unsigned*)reloc) += sym_addr;
1399            break;
1400        case R_ARM_REL32:
1401            COUNT_RELOC(RELOC_RELATIVE);
1402            MARK(rel->r_offset);
1403            TRACE_TYPE(RELO, "%5d RELO REL32 %08x <- %08x - %08x %s\n", pid,
1404                       reloc, sym_addr, rel->r_offset, sym_name);
1405            *((unsigned*)reloc) += sym_addr - rel->r_offset;
1406            break;
1407#elif defined(ANDROID_X86_LINKER)
1408        case R_386_JUMP_SLOT:
1409            COUNT_RELOC(RELOC_ABSOLUTE);
1410            MARK(rel->r_offset);
1411            TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", pid,
1412                       reloc, sym_addr, sym_name);
1413            *((unsigned*)reloc) = sym_addr;
1414            break;
1415        case R_386_GLOB_DAT:
1416            COUNT_RELOC(RELOC_ABSOLUTE);
1417            MARK(rel->r_offset);
1418            TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", pid,
1419                       reloc, sym_addr, sym_name);
1420            *((unsigned*)reloc) = sym_addr;
1421            break;
1422#endif /* ANDROID_*_LINKER */
1423
1424#if defined(ANDROID_ARM_LINKER)
1425        case R_ARM_RELATIVE:
1426#elif defined(ANDROID_X86_LINKER)
1427        case R_386_RELATIVE:
1428#endif /* ANDROID_*_LINKER */
1429            COUNT_RELOC(RELOC_RELATIVE);
1430            MARK(rel->r_offset);
1431            if(sym){
1432                DL_ERR("%5d odd RELATIVE form...", pid);
1433                return -1;
1434            }
1435            TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", pid,
1436                       reloc, si->base);
1437            *((unsigned*)reloc) += si->base;
1438            break;
1439
1440#if defined(ANDROID_X86_LINKER)
1441        case R_386_32:
1442            COUNT_RELOC(RELOC_RELATIVE);
1443            MARK(rel->r_offset);
1444
1445            TRACE_TYPE(RELO, "%5d RELO R_386_32 %08x <- +%08x %s\n", pid,
1446                       reloc, sym_addr, sym_name);
1447            *((unsigned *)reloc) += (unsigned)sym_addr;
1448            break;
1449
1450        case R_386_PC32:
1451            COUNT_RELOC(RELOC_RELATIVE);
1452            MARK(rel->r_offset);
1453            TRACE_TYPE(RELO, "%5d RELO R_386_PC32 %08x <- "
1454                       "+%08x (%08x - %08x) %s\n", pid, reloc,
1455                       (sym_addr - reloc), sym_addr, reloc, sym_name);
1456            *((unsigned *)reloc) += (unsigned)(sym_addr - reloc);
1457            break;
1458#endif /* ANDROID_X86_LINKER */
1459
1460#ifdef ANDROID_ARM_LINKER
1461        case R_ARM_COPY:
1462            COUNT_RELOC(RELOC_COPY);
1463            MARK(rel->r_offset);
1464            TRACE_TYPE(RELO, "%5d RELO %08x <- %d @ %08x %s\n", pid,
1465                       reloc, s->st_size, sym_addr, sym_name);
1466            memcpy((void*)reloc, (void*)sym_addr, s->st_size);
1467            break;
1468        case R_ARM_NONE:
1469            break;
1470#endif /* ANDROID_ARM_LINKER */
1471
1472        default:
1473            DL_ERR("%5d unknown reloc type %d @ %p (%d)",
1474                  pid, type, rel, (int) (rel - start));
1475            return -1;
1476        }
1477        rel++;
1478    }
1479    return 0;
1480}
1481
1482/* Please read the "Initialization and Termination functions" functions.
1483 * of the linker design note in bionic/linker/README.TXT to understand
1484 * what the following code is doing.
1485 *
1486 * The important things to remember are:
1487 *
1488 *   DT_PREINIT_ARRAY must be called first for executables, and should
1489 *   not appear in shared libraries.
1490 *
1491 *   DT_INIT should be called before DT_INIT_ARRAY if both are present
1492 *
1493 *   DT_FINI should be called after DT_FINI_ARRAY if both are present
1494 *
1495 *   DT_FINI_ARRAY must be parsed in reverse order.
1496 */
1497
1498static void call_array(unsigned *ctor, int count, int reverse)
1499{
1500    int n, inc = 1;
1501
1502    if (reverse) {
1503        ctor += (count-1);
1504        inc   = -1;
1505    }
1506
1507    for(n = count; n > 0; n--) {
1508        TRACE("[ %5d Looking at %s *0x%08x == 0x%08x ]\n", pid,
1509              reverse ? "dtor" : "ctor",
1510              (unsigned)ctor, (unsigned)*ctor);
1511        void (*func)() = (void (*)()) *ctor;
1512        ctor += inc;
1513        if(((int) func == 0) || ((int) func == -1)) continue;
1514        TRACE("[ %5d Calling func @ 0x%08x ]\n", pid, (unsigned)func);
1515        func();
1516    }
1517}
1518
1519void call_constructors_recursive(soinfo *si)
1520{
1521    if (si->constructors_called)
1522        return;
1523
1524    // Set this before actually calling the constructors, otherwise it doesn't
1525    // protect against recursive constructor calls. One simple example of
1526    // constructor recursion is the libc debug malloc, which is implemented in
1527    // libc_malloc_debug_leak.so:
1528    // 1. The program depends on libc, so libc's constructor is called here.
1529    // 2. The libc constructor calls dlopen() to load libc_malloc_debug_leak.so.
1530    // 3. dlopen() calls call_constructors_recursive() with the newly created
1531    //    soinfo for libc_malloc_debug_leak.so.
1532    // 4. The debug so depends on libc, so call_constructors_recursive() is
1533    //    called again with the libc soinfo. If it doesn't trigger the early-
1534    //    out above, the libc constructor will be called again (recursively!).
1535    si->constructors_called = 1;
1536
1537    if (si->flags & FLAG_EXE) {
1538        TRACE("[ %5d Calling preinit_array @ 0x%08x [%d] for '%s' ]\n",
1539              pid, (unsigned)si->preinit_array, si->preinit_array_count,
1540              si->name);
1541        call_array(si->preinit_array, si->preinit_array_count, 0);
1542        TRACE("[ %5d Done calling preinit_array for '%s' ]\n", pid, si->name);
1543    } else {
1544        if (si->preinit_array) {
1545            DL_ERR("%5d Shared library '%s' has a preinit_array table @ 0x%08x."
1546                   " This is INVALID.", pid, si->name,
1547                   (unsigned)si->preinit_array);
1548        }
1549    }
1550
1551    if (si->dynamic) {
1552        unsigned *d;
1553        for(d = si->dynamic; *d; d += 2) {
1554            if(d[0] == DT_NEEDED){
1555                soinfo* lsi = (soinfo *)d[1];
1556                if (!validate_soinfo(lsi)) {
1557                    DL_ERR("%5d bad DT_NEEDED pointer in %s",
1558                           pid, si->name);
1559                } else {
1560                    call_constructors_recursive(lsi);
1561                }
1562            }
1563        }
1564    }
1565
1566    if (si->init_func) {
1567        TRACE("[ %5d Calling init_func @ 0x%08x for '%s' ]\n", pid,
1568              (unsigned)si->init_func, si->name);
1569        si->init_func();
1570        TRACE("[ %5d Done calling init_func for '%s' ]\n", pid, si->name);
1571    }
1572
1573    if (si->init_array) {
1574        TRACE("[ %5d Calling init_array @ 0x%08x [%d] for '%s' ]\n", pid,
1575              (unsigned)si->init_array, si->init_array_count, si->name);
1576        call_array(si->init_array, si->init_array_count, 0);
1577        TRACE("[ %5d Done calling init_array for '%s' ]\n", pid, si->name);
1578    }
1579
1580}
1581
1582static void call_destructors(soinfo *si)
1583{
1584    if (si->fini_array) {
1585        TRACE("[ %5d Calling fini_array @ 0x%08x [%d] for '%s' ]\n", pid,
1586              (unsigned)si->fini_array, si->fini_array_count, si->name);
1587        call_array(si->fini_array, si->fini_array_count, 1);
1588        TRACE("[ %5d Done calling fini_array for '%s' ]\n", pid, si->name);
1589    }
1590
1591    if (si->fini_func) {
1592        TRACE("[ %5d Calling fini_func @ 0x%08x for '%s' ]\n", pid,
1593              (unsigned)si->fini_func, si->name);
1594        si->fini_func();
1595        TRACE("[ %5d Done calling fini_func for '%s' ]\n", pid, si->name);
1596    }
1597}
1598
1599/* Force any of the closed stdin, stdout and stderr to be associated with
1600   /dev/null. */
1601static int nullify_closed_stdio (void)
1602{
1603    int dev_null, i, status;
1604    int return_value = 0;
1605
1606    dev_null = open("/dev/null", O_RDWR);
1607    if (dev_null < 0) {
1608        DL_ERR("Cannot open /dev/null.");
1609        return -1;
1610    }
1611    TRACE("[ %5d Opened /dev/null file-descriptor=%d]\n", pid, dev_null);
1612
1613    /* If any of the stdio file descriptors is valid and not associated
1614       with /dev/null, dup /dev/null to it.  */
1615    for (i = 0; i < 3; i++) {
1616        /* If it is /dev/null already, we are done. */
1617        if (i == dev_null)
1618            continue;
1619
1620        TRACE("[ %5d Nullifying stdio file descriptor %d]\n", pid, i);
1621        /* The man page of fcntl does not say that fcntl(..,F_GETFL)
1622           can be interrupted but we do this just to be safe. */
1623        do {
1624          status = fcntl(i, F_GETFL);
1625        } while (status < 0 && errno == EINTR);
1626
1627        /* If file is openned, we are good. */
1628        if (status >= 0)
1629          continue;
1630
1631        /* The only error we allow is that the file descriptor does not
1632           exist, in which case we dup /dev/null to it. */
1633        if (errno != EBADF) {
1634            DL_ERR("nullify_stdio: unhandled error %s", strerror(errno));
1635            return_value = -1;
1636            continue;
1637        }
1638
1639        /* Try dupping /dev/null to this stdio file descriptor and
1640           repeat if there is a signal.  Note that any errors in closing
1641           the stdio descriptor are lost.  */
1642        do {
1643            status = dup2(dev_null, i);
1644        } while (status < 0 && errno == EINTR);
1645
1646        if (status < 0) {
1647            DL_ERR("nullify_stdio: dup2 error %s", strerror(errno));
1648            return_value = -1;
1649            continue;
1650        }
1651    }
1652
1653    /* If /dev/null is not one of the stdio file descriptors, close it. */
1654    if (dev_null > 2) {
1655        TRACE("[ %5d Closing /dev/null file-descriptor=%d]\n", pid, dev_null);
1656        do {
1657            status = close(dev_null);
1658        } while (status < 0 && errno == EINTR);
1659
1660        if (status < 0) {
1661            DL_ERR("nullify_stdio: close error %s", strerror(errno));
1662            return_value = -1;
1663        }
1664    }
1665
1666    return return_value;
1667}
1668
1669static int link_image(soinfo *si, unsigned wr_offset)
1670{
1671    unsigned *d;
1672    Elf32_Phdr *phdr = si->phdr;
1673    int phnum = si->phnum;
1674
1675    INFO("[ %5d linking %s ]\n", pid, si->name);
1676    DEBUG("%5d si->base = 0x%08x si->flags = 0x%08x\n", pid,
1677          si->base, si->flags);
1678
1679    if (si->flags & (FLAG_EXE | FLAG_LINKER)) {
1680        /* Locate the needed program segments (DYNAMIC/ARM_EXIDX) for
1681         * linkage info if this is the executable or the linker itself.
1682         * If this was a dynamic lib, that would have been done at load time.
1683         *
1684         * TODO: It's unfortunate that small pieces of this are
1685         * repeated from the load_library routine. Refactor this just
1686         * slightly to reuse these bits.
1687         */
1688        si->size = 0;
1689        for(; phnum > 0; --phnum, ++phdr) {
1690#ifdef ANDROID_ARM_LINKER
1691            if(phdr->p_type == PT_ARM_EXIDX) {
1692                /* exidx entries (used for stack unwinding) are 8 bytes each.
1693                 */
1694                si->ARM_exidx = (unsigned *)phdr->p_vaddr;
1695                si->ARM_exidx_count = phdr->p_memsz / 8;
1696            }
1697#endif
1698            if (phdr->p_type == PT_LOAD) {
1699                /* For the executable, we use the si->size field only in
1700                   dl_unwind_find_exidx(), so the meaning of si->size
1701                   is not the size of the executable; it is the distance
1702                   between the load location of the executable and the last
1703                   address of the loadable part of the executable.
1704                   We use the range [si->base, si->base + si->size) to
1705                   determine whether a PC value falls within the executable
1706                   section. Of course, if a value is between si->base and
1707                   (si->base + phdr->p_vaddr), it's not in the executable
1708                   section, but a) we shouldn't be asking for such a value
1709                   anyway, and b) if we have to provide an EXIDX for such a
1710                   value, then the executable's EXIDX is probably the better
1711                   choice.
1712                */
1713                DEBUG_DUMP_PHDR(phdr, "PT_LOAD", pid);
1714                if (phdr->p_vaddr + phdr->p_memsz > si->size)
1715                    si->size = phdr->p_vaddr + phdr->p_memsz;
1716                /* try to remember what range of addresses should be write
1717                 * protected */
1718                if (!(phdr->p_flags & PF_W)) {
1719                    unsigned _end;
1720
1721                    if (si->base + phdr->p_vaddr < si->wrprotect_start)
1722                        si->wrprotect_start = si->base + phdr->p_vaddr;
1723                    _end = (((si->base + phdr->p_vaddr + phdr->p_memsz + PAGE_SIZE - 1) &
1724                             (~PAGE_MASK)));
1725                    if (_end > si->wrprotect_end)
1726                        si->wrprotect_end = _end;
1727                    /* Make the section writable just in case we'll have to
1728                     * write to it during relocation (i.e. text segment).
1729                     * However, we will remember what range of addresses
1730                     * should be write protected.
1731                     */
1732                    mprotect((void *) (si->base + phdr->p_vaddr),
1733                             phdr->p_memsz,
1734                             PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
1735                }
1736            } else if (phdr->p_type == PT_DYNAMIC) {
1737                if (si->dynamic != (unsigned *)-1) {
1738                    DL_ERR("%5d multiple PT_DYNAMIC segments found in '%s'. "
1739                          "Segment at 0x%08x, previously one found at 0x%08x",
1740                          pid, si->name, si->base + phdr->p_vaddr,
1741                          (unsigned)si->dynamic);
1742                    goto fail;
1743                }
1744                DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", pid);
1745                si->dynamic = (unsigned *) (si->base + phdr->p_vaddr);
1746            } else if (phdr->p_type == PT_GNU_RELRO) {
1747                if ((phdr->p_vaddr >= si->size)
1748                        || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
1749                        || ((si->base + phdr->p_vaddr + phdr->p_memsz) < si->base)) {
1750                    DL_ERR("%d invalid GNU_RELRO in '%s' "
1751                           "p_vaddr=0x%08x p_memsz=0x%08x", pid, si->name,
1752                           phdr->p_vaddr, phdr->p_memsz);
1753                    goto fail;
1754                }
1755                si->gnu_relro_start = (Elf32_Addr) (si->base + phdr->p_vaddr);
1756                si->gnu_relro_len = (unsigned) phdr->p_memsz;
1757            }
1758        }
1759    }
1760
1761    if (si->dynamic == (unsigned *)-1) {
1762        DL_ERR("%5d missing PT_DYNAMIC?!", pid);
1763        goto fail;
1764    }
1765
1766    DEBUG("%5d dynamic = %p\n", pid, si->dynamic);
1767
1768    /* extract useful information from dynamic section */
1769    for(d = si->dynamic; *d; d++){
1770        DEBUG("%5d d = %p, d[0] = 0x%08x d[1] = 0x%08x\n", pid, d, d[0], d[1]);
1771        switch(*d++){
1772        case DT_HASH:
1773            si->nbucket = ((unsigned *) (si->base + *d))[0];
1774            si->nchain = ((unsigned *) (si->base + *d))[1];
1775            si->bucket = (unsigned *) (si->base + *d + 8);
1776            si->chain = (unsigned *) (si->base + *d + 8 + si->nbucket * 4);
1777            break;
1778        case DT_STRTAB:
1779            si->strtab = (const char *) (si->base + *d);
1780            break;
1781        case DT_SYMTAB:
1782            si->symtab = (Elf32_Sym *) (si->base + *d);
1783            break;
1784        case DT_PLTREL:
1785            if(*d != DT_REL) {
1786                DL_ERR("DT_RELA not supported");
1787                goto fail;
1788            }
1789            break;
1790        case DT_JMPREL:
1791            si->plt_rel = (Elf32_Rel*) (si->base + *d);
1792            break;
1793        case DT_PLTRELSZ:
1794            si->plt_rel_count = *d / 8;
1795            break;
1796        case DT_REL:
1797            si->rel = (Elf32_Rel*) (si->base + *d);
1798            break;
1799        case DT_RELSZ:
1800            si->rel_count = *d / 8;
1801            break;
1802        case DT_PLTGOT:
1803            /* Save this in case we decide to do lazy binding. We don't yet. */
1804            si->plt_got = (unsigned *)(si->base + *d);
1805            break;
1806        case DT_DEBUG:
1807            // Set the DT_DEBUG entry to the addres of _r_debug for GDB
1808            *d = (int) &_r_debug;
1809            break;
1810         case DT_RELA:
1811            DL_ERR("%5d DT_RELA not supported", pid);
1812            goto fail;
1813        case DT_INIT:
1814            si->init_func = (void (*)(void))(si->base + *d);
1815            DEBUG("%5d %s constructors (init func) found at %p\n",
1816                  pid, si->name, si->init_func);
1817            break;
1818        case DT_FINI:
1819            si->fini_func = (void (*)(void))(si->base + *d);
1820            DEBUG("%5d %s destructors (fini func) found at %p\n",
1821                  pid, si->name, si->fini_func);
1822            break;
1823        case DT_INIT_ARRAY:
1824            si->init_array = (unsigned *)(si->base + *d);
1825            DEBUG("%5d %s constructors (init_array) found at %p\n",
1826                  pid, si->name, si->init_array);
1827            break;
1828        case DT_INIT_ARRAYSZ:
1829            si->init_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
1830            break;
1831        case DT_FINI_ARRAY:
1832            si->fini_array = (unsigned *)(si->base + *d);
1833            DEBUG("%5d %s destructors (fini_array) found at %p\n",
1834                  pid, si->name, si->fini_array);
1835            break;
1836        case DT_FINI_ARRAYSZ:
1837            si->fini_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
1838            break;
1839        case DT_PREINIT_ARRAY:
1840            si->preinit_array = (unsigned *)(si->base + *d);
1841            DEBUG("%5d %s constructors (preinit_array) found at %p\n",
1842                  pid, si->name, si->preinit_array);
1843            break;
1844        case DT_PREINIT_ARRAYSZ:
1845            si->preinit_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
1846            break;
1847        case DT_TEXTREL:
1848            /* TODO: make use of this. */
1849            /* this means that we might have to write into where the text
1850             * segment was loaded during relocation... Do something with
1851             * it.
1852             */
1853            DEBUG("%5d Text segment should be writable during relocation.\n",
1854                  pid);
1855            break;
1856        }
1857    }
1858
1859    DEBUG("%5d si->base = 0x%08x, si->strtab = %p, si->symtab = %p\n",
1860           pid, si->base, si->strtab, si->symtab);
1861
1862    if((si->strtab == 0) || (si->symtab == 0)) {
1863        DL_ERR("%5d missing essential tables", pid);
1864        goto fail;
1865    }
1866
1867    /* if this is the main executable, then load all of the preloads now */
1868    if(si->flags & FLAG_EXE) {
1869        int i;
1870        memset(preloads, 0, sizeof(preloads));
1871        for(i = 0; ldpreload_names[i] != NULL; i++) {
1872            soinfo *lsi = find_library(ldpreload_names[i]);
1873            if(lsi == 0) {
1874                strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
1875                DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
1876                       pid, ldpreload_names[i], si->name, tmp_err_buf);
1877                goto fail;
1878            }
1879            lsi->refcount++;
1880            preloads[i] = lsi;
1881        }
1882    }
1883
1884    for(d = si->dynamic; *d; d += 2) {
1885        if(d[0] == DT_NEEDED){
1886            DEBUG("%5d %s needs %s\n", pid, si->name, si->strtab + d[1]);
1887            soinfo *lsi = find_library(si->strtab + d[1]);
1888            if(lsi == 0) {
1889                strlcpy(tmp_err_buf, linker_get_error(), sizeof(tmp_err_buf));
1890                DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
1891                       pid, si->strtab + d[1], si->name, tmp_err_buf);
1892                goto fail;
1893            }
1894            /* Save the soinfo of the loaded DT_NEEDED library in the payload
1895               of the DT_NEEDED entry itself, so that we can retrieve the
1896               soinfo directly later from the dynamic segment.  This is a hack,
1897               but it allows us to map from DT_NEEDED to soinfo efficiently
1898               later on when we resolve relocations, trying to look up a symbol
1899               with dlsym().
1900            */
1901            d[1] = (unsigned)lsi;
1902            lsi->refcount++;
1903        }
1904    }
1905
1906    if(si->plt_rel) {
1907        DEBUG("[ %5d relocating %s plt ]\n", pid, si->name );
1908        if(reloc_library(si, si->plt_rel, si->plt_rel_count))
1909            goto fail;
1910    }
1911    if(si->rel) {
1912        DEBUG("[ %5d relocating %s ]\n", pid, si->name );
1913        if(reloc_library(si, si->rel, si->rel_count))
1914            goto fail;
1915    }
1916
1917    si->flags |= FLAG_LINKED;
1918    DEBUG("[ %5d finished linking %s ]\n", pid, si->name);
1919
1920#if 0
1921    /* This is the way that the old dynamic linker did protection of
1922     * non-writable areas. It would scan section headers and find where
1923     * .text ended (rather where .data/.bss began) and assume that this is
1924     * the upper range of the non-writable area. This is too coarse,
1925     * and is kept here for reference until we fully move away from single
1926     * segment elf objects. See the code in get_wr_offset (also #if'd 0)
1927     * that made this possible.
1928     */
1929    if(wr_offset < 0xffffffff){
1930        mprotect((void*) si->base, wr_offset, PROT_READ | PROT_EXEC);
1931    }
1932#else
1933    /* TODO: Verify that this does the right thing in all cases, as it
1934     * presently probably does not. It is possible that an ELF image will
1935     * come with multiple read-only segments. What we ought to do is scan
1936     * the program headers again and mprotect all the read-only segments.
1937     * To prevent re-scanning the program header, we would have to build a
1938     * list of loadable segments in si, and then scan that instead. */
1939    if (si->wrprotect_start != 0xffffffff && si->wrprotect_end != 0) {
1940        mprotect((void *)si->wrprotect_start,
1941                 si->wrprotect_end - si->wrprotect_start,
1942                 PROT_READ | PROT_EXEC);
1943    }
1944#endif
1945
1946    if (si->gnu_relro_start != 0 && si->gnu_relro_len != 0) {
1947        Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
1948        unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
1949        if (mprotect((void *) start, len, PROT_READ) < 0) {
1950            DL_ERR("%5d GNU_RELRO mprotect of library '%s' failed: %d (%s)\n",
1951                   pid, si->name, errno, strerror(errno));
1952            goto fail;
1953        }
1954    }
1955
1956    /* If this is a SET?ID program, dup /dev/null to opened stdin,
1957       stdout and stderr to close a security hole described in:
1958
1959    ftp://ftp.freebsd.org/pub/FreeBSD/CERT/advisories/FreeBSD-SA-02:23.stdio.asc
1960
1961     */
1962    if (program_is_setuid)
1963        nullify_closed_stdio ();
1964    notify_gdb_of_load(si);
1965    return 0;
1966
1967fail:
1968    ERROR("failed to link %s\n", si->name);
1969    si->flags |= FLAG_ERROR;
1970    return -1;
1971}
1972
1973static void parse_library_path(const char *path, char *delim)
1974{
1975    size_t len;
1976    char *ldpaths_bufp = ldpaths_buf;
1977    int i = 0;
1978
1979    len = strlcpy(ldpaths_buf, path, sizeof(ldpaths_buf));
1980
1981    while (i < LDPATH_MAX && (ldpaths[i] = strsep(&ldpaths_bufp, delim))) {
1982        if (*ldpaths[i] != '\0')
1983            ++i;
1984    }
1985
1986    /* Forget the last path if we had to truncate; this occurs if the 2nd to
1987     * last char isn't '\0' (i.e. not originally a delim). */
1988    if (i > 0 && len >= sizeof(ldpaths_buf) &&
1989            ldpaths_buf[sizeof(ldpaths_buf) - 2] != '\0') {
1990        ldpaths[i - 1] = NULL;
1991    } else {
1992        ldpaths[i] = NULL;
1993    }
1994}
1995
1996static void parse_preloads(const char *path, char *delim)
1997{
1998    size_t len;
1999    char *ldpreloads_bufp = ldpreloads_buf;
2000    int i = 0;
2001
2002    len = strlcpy(ldpreloads_buf, path, sizeof(ldpreloads_buf));
2003
2004    while (i < LDPRELOAD_MAX && (ldpreload_names[i] = strsep(&ldpreloads_bufp, delim))) {
2005        if (*ldpreload_names[i] != '\0') {
2006            ++i;
2007        }
2008    }
2009
2010    /* Forget the last path if we had to truncate; this occurs if the 2nd to
2011     * last char isn't '\0' (i.e. not originally a delim). */
2012    if (i > 0 && len >= sizeof(ldpreloads_buf) &&
2013            ldpreloads_buf[sizeof(ldpreloads_buf) - 2] != '\0') {
2014        ldpreload_names[i - 1] = NULL;
2015    } else {
2016        ldpreload_names[i] = NULL;
2017    }
2018}
2019
2020/*
2021 * This code is called after the linker has linked itself and
2022 * fixed it's own GOT. It is safe to make references to externs
2023 * and other non-local data at this point.
2024 */
2025static unsigned __linker_init_post_relocation(unsigned **elfdata)
2026{
2027    static soinfo linker_soinfo;
2028
2029    int argc = (int) *elfdata;
2030    char **argv = (char**) (elfdata + 1);
2031    unsigned *vecs = (unsigned*) (argv + argc + 1);
2032    unsigned *v;
2033    soinfo *si;
2034    struct link_map * map;
2035    const char *ldpath_env = NULL;
2036    const char *ldpreload_env = NULL;
2037
2038    /* NOTE: we store the elfdata pointer on a special location
2039     *       of the temporary TLS area in order to pass it to
2040     *       the C Library's runtime initializer.
2041     *
2042     *       The initializer must clear the slot and reset the TLS
2043     *       to point to a different location to ensure that no other
2044     *       shared library constructor can access it.
2045     */
2046    __libc_init_tls(elfdata);
2047
2048    pid = getpid();
2049
2050#if TIMING
2051    struct timeval t0, t1;
2052    gettimeofday(&t0, 0);
2053#endif
2054
2055    /* Initialize environment functions, and get to the ELF aux vectors table */
2056    vecs = linker_env_init(vecs);
2057
2058    /* Check auxv for AT_SECURE first to see if program is setuid, setgid,
2059       has file caps, or caused a SELinux/AppArmor domain transition. */
2060    for (v = vecs; v[0]; v += 2) {
2061        if (v[0] == AT_SECURE) {
2062            /* kernel told us whether to enable secure mode */
2063            program_is_setuid = v[1];
2064            goto sanitize;
2065        }
2066    }
2067
2068    /* Kernel did not provide AT_SECURE - fall back on legacy test. */
2069    program_is_setuid = (getuid() != geteuid()) || (getgid() != getegid());
2070
2071sanitize:
2072    /* Sanitize environment if we're loading a setuid program */
2073    if (program_is_setuid)
2074        linker_env_secure();
2075
2076    debugger_init();
2077
2078    /* Get a few environment variables */
2079    {
2080#if LINKER_DEBUG
2081        const char* env;
2082        env = linker_env_get("DEBUG"); /* XXX: TODO: Change to LD_DEBUG */
2083        if (env)
2084            debug_verbosity = atoi(env);
2085#endif
2086
2087        /* Normally, these are cleaned by linker_env_secure, but the test
2088         * against program_is_setuid doesn't cost us anything */
2089        if (!program_is_setuid) {
2090            ldpath_env = linker_env_get("LD_LIBRARY_PATH");
2091            ldpreload_env = linker_env_get("LD_PRELOAD");
2092        }
2093    }
2094
2095    INFO("[ android linker & debugger ]\n");
2096    DEBUG("%5d elfdata @ 0x%08x\n", pid, (unsigned)elfdata);
2097
2098    si = alloc_info(argv[0]);
2099    if(si == 0) {
2100        exit(-1);
2101    }
2102
2103        /* bootstrap the link map, the main exe always needs to be first */
2104    si->flags |= FLAG_EXE;
2105    map = &(si->linkmap);
2106
2107    map->l_addr = 0;
2108    map->l_name = argv[0];
2109    map->l_prev = NULL;
2110    map->l_next = NULL;
2111
2112    _r_debug.r_map = map;
2113    r_debug_tail = map;
2114
2115        /* gdb expects the linker to be in the debug shared object list,
2116         * and we need to make sure that the reported load address is zero.
2117         * Without this, gdb gets the wrong idea of where rtld_db_dlactivity()
2118         * is.  Don't use alloc_info(), because the linker shouldn't
2119         * be on the soinfo list.
2120         */
2121    strlcpy((char*) linker_soinfo.name, "/system/bin/linker", sizeof linker_soinfo.name);
2122    linker_soinfo.flags = 0;
2123    linker_soinfo.base = 0;     // This is the important part; must be zero.
2124    insert_soinfo_into_debug_map(&linker_soinfo);
2125
2126        /* extract information passed from the kernel */
2127    while(vecs[0] != 0){
2128        switch(vecs[0]){
2129        case AT_PHDR:
2130            si->phdr = (Elf32_Phdr*) vecs[1];
2131            break;
2132        case AT_PHNUM:
2133            si->phnum = (int) vecs[1];
2134            break;
2135        case AT_ENTRY:
2136            si->entry = vecs[1];
2137            break;
2138        }
2139        vecs += 2;
2140    }
2141
2142    /* Compute the value of si->base. We can't rely on the fact that
2143     * the first entry is the PHDR because this will not be true
2144     * for certain executables (e.g. some in the NDK unit test suite)
2145     */
2146    int nn;
2147    si->base = 0;
2148    for ( nn = 0; nn < si->phnum; nn++ ) {
2149        if (si->phdr[nn].p_type == PT_PHDR) {
2150            si->base = (Elf32_Addr) si->phdr - si->phdr[nn].p_vaddr;
2151            break;
2152        }
2153    }
2154    si->dynamic = (unsigned *)-1;
2155    si->wrprotect_start = 0xffffffff;
2156    si->wrprotect_end = 0;
2157    si->refcount = 1;
2158    si->gnu_relro_start = 0;
2159    si->gnu_relro_len = 0;
2160
2161        /* Use LD_LIBRARY_PATH if we aren't setuid/setgid */
2162    if (ldpath_env)
2163        parse_library_path(ldpath_env, ":");
2164
2165    if (ldpreload_env) {
2166        parse_preloads(ldpreload_env, " :");
2167    }
2168
2169    if(link_image(si, 0)) {
2170        char errmsg[] = "CANNOT LINK EXECUTABLE\n";
2171        write(2, __linker_dl_err_buf, strlen(__linker_dl_err_buf));
2172        write(2, errmsg, sizeof(errmsg));
2173        exit(-1);
2174    }
2175
2176    call_constructors_recursive(si);
2177
2178#if ALLOW_SYMBOLS_FROM_MAIN
2179    /* Set somain after we've loaded all the libraries in order to prevent
2180     * linking of symbols back to the main image, which is not set up at that
2181     * point yet.
2182     */
2183    somain = si;
2184#endif
2185
2186#if TIMING
2187    gettimeofday(&t1,NULL);
2188    PRINT("LINKER TIME: %s: %d microseconds\n", argv[0], (int) (
2189               (((long long)t1.tv_sec * 1000000LL) + (long long)t1.tv_usec) -
2190               (((long long)t0.tv_sec * 1000000LL) + (long long)t0.tv_usec)
2191               ));
2192#endif
2193#if STATS
2194    PRINT("RELO STATS: %s: %d abs, %d rel, %d copy, %d symbol\n", argv[0],
2195           linker_stats.reloc[RELOC_ABSOLUTE],
2196           linker_stats.reloc[RELOC_RELATIVE],
2197           linker_stats.reloc[RELOC_COPY],
2198           linker_stats.reloc[RELOC_SYMBOL]);
2199#endif
2200#if COUNT_PAGES
2201    {
2202        unsigned n;
2203        unsigned i;
2204        unsigned count = 0;
2205        for(n = 0; n < 4096; n++){
2206            if(bitmask[n]){
2207                unsigned x = bitmask[n];
2208                for(i = 0; i < 8; i++){
2209                    if(x & 1) count++;
2210                    x >>= 1;
2211                }
2212            }
2213        }
2214        PRINT("PAGES MODIFIED: %s: %d (%dKB)\n", argv[0], count, count * 4);
2215    }
2216#endif
2217
2218#if TIMING || STATS || COUNT_PAGES
2219    fflush(stdout);
2220#endif
2221
2222    TRACE("[ %5d Ready to execute '%s' @ 0x%08x ]\n", pid, si->name,
2223          si->entry);
2224    return si->entry;
2225}
2226
2227/*
2228 * Find the value of AT_BASE passed to us by the kernel. This is the load
2229 * location of the linker.
2230 */
2231static unsigned find_linker_base(unsigned **elfdata) {
2232    int argc = (int) *elfdata;
2233    char **argv = (char**) (elfdata + 1);
2234    unsigned *vecs = (unsigned*) (argv + argc + 1);
2235    while (vecs[0] != 0) {
2236        vecs++;
2237    }
2238
2239    /* The end of the environment block is marked by two NULL pointers */
2240    vecs++;
2241
2242    while(vecs[0]) {
2243        if (vecs[0] == AT_BASE) {
2244            return vecs[1];
2245        }
2246        vecs += 2;
2247    }
2248
2249    return 0; // should never happen
2250}
2251
2252/*
2253 * This is the entry point for the linker, called from begin.S. This
2254 * method is responsible for fixing the linker's own relocations, and
2255 * then calling __linker_init_post_relocation().
2256 *
2257 * Because this method is called before the linker has fixed it's own
2258 * relocations, any attempt to reference an extern variable, extern
2259 * function, or other GOT reference will generate a segfault.
2260 */
2261unsigned __linker_init(unsigned **elfdata) {
2262    unsigned linker_addr = find_linker_base(elfdata);
2263    Elf32_Ehdr *elf_hdr = (Elf32_Ehdr *) linker_addr;
2264    Elf32_Phdr *phdr =
2265        (Elf32_Phdr *)((unsigned char *) linker_addr + elf_hdr->e_phoff);
2266
2267    soinfo linker_so;
2268    memset(&linker_so, 0, sizeof(soinfo));
2269
2270    linker_so.base = linker_addr;
2271    linker_so.dynamic = (unsigned *) -1;
2272    linker_so.phdr = phdr;
2273    linker_so.phnum = elf_hdr->e_phnum;
2274    linker_so.flags |= FLAG_LINKER;
2275    linker_so.wrprotect_start = 0xffffffff;
2276    linker_so.wrprotect_end = 0;
2277    linker_so.gnu_relro_start = 0;
2278    linker_so.gnu_relro_len = 0;
2279
2280    if (link_image(&linker_so, 0)) {
2281        // It would be nice to print an error message, but if the linker
2282        // can't link itself, there's no guarantee that we'll be able to
2283        // call write() (because it involves a GOT reference).
2284        //
2285        // This situation should never occur unless the linker itself
2286        // is corrupt.
2287        exit(-1);
2288    }
2289
2290    // We have successfully fixed our own relocations. It's safe to run
2291    // the main part of the linker now.
2292    return __linker_init_post_relocation(elfdata);
2293}
2294