summaryrefslogtreecommitdiff
path: root/src/linker/linker.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/linker/linker.c')
-rw-r--r--src/linker/linker.c2541
1 files changed, 2541 insertions, 0 deletions
diff --git a/src/linker/linker.c b/src/linker/linker.c
new file mode 100644
index 0000000..8f95472
--- /dev/null
+++ b/src/linker/linker.c
@@ -0,0 +1,2541 @@
+/*
+ * Copyright (C) 2008, 2009 The Android Open Source Project
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
+ * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
+ * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
+ * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
+ * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
+ * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
+ * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
+ * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+#ifdef APKENV_DEBUG
+# define LINKER_DEBUG_PRINTF(...) PRINT(__VA_ARGS__)
+#else
+# define LINKER_DEBUG_PRINTF(...)
+#endif
+
+
+#include <linux/auxvec.h>
+
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+#include <unistd.h>
+#include <fcntl.h>
+#include <errno.h>
+#include <dlfcn.h>
+#include <sys/stat.h>
+#include <assert.h>
+
+#include <pthread.h>
+
+#include <sys/mman.h>
+
+/* special private C library header - see Android.mk */
+#include "bionic_tls.h"
+
+#include "strlcpy.h"
+
+/* for apkenv_get_hooked_symbol */
+// #include "../compat/hooks.h"
+/* for create_wrapper */
+// #include "../debug/wrappers.h"
+
+#include "linker.h"
+#include "linker_debug.h"
+#include "linker_environ.h"
+#include "linker_format.h"
+
+#define ALLOW_SYMBOLS_FROM_MAIN 1
+#define SO_MAX 128
+
+/* Assume average path length of 64 and max 8 paths */
+#define LDPATH_BUFSIZE 512
+#define LDPATH_MAX 8
+
+#define LDPRELOAD_BUFSIZE 512
+#define LDPRELOAD_MAX 8
+
+/* >>> IMPORTANT NOTE - READ ME BEFORE MODIFYING <<<
+ *
+ * Do NOT use malloc() and friends or pthread_*() code here.
+ * Don't use printf() either; it's caused mysterious memory
+ * corruption in the past.
+ * The linker runs before we bring up libc and it's easiest
+ * to make sure it does not depend on any complex libc features
+ *
+ * open issues / todo:
+ *
+ * - are we doing everything we should for ARM_COPY relocations?
+ * - cleaner error reporting
+ * - after linking, set as much stuff as possible to READONLY
+ * and NOEXEC
+ * - linker hardcodes PAGE_SIZE and PAGE_MASK because the kernel
+ * headers provide versions that are negative...
+ * - allocate space for soinfo structs dynamically instead of
+ * having a hard limit (64)
+*/
+
+
+static int apkenv_link_image(soinfo *si, unsigned wr_offset);
+
+static int apkenv_socount = 0;
+static soinfo apkenv_sopool[SO_MAX];
+static soinfo *apkenv_freelist = NULL;
+static soinfo *apkenv_solist = &apkenv_libdl_info;
+static soinfo *apkenv_sonext = &apkenv_libdl_info;
+#if ALLOW_SYMBOLS_FROM_MAIN
+static soinfo *apkenv_somain; /* main process, always the one after apkenv_libdl_info */
+#endif
+
+
+static inline int apkenv_validate_soinfo(soinfo *si)
+{
+ return (si >= apkenv_sopool && si < apkenv_sopool + SO_MAX) ||
+ si == &apkenv_libdl_info;
+}
+
+static char apkenv_ldpaths_buf[LDPATH_BUFSIZE];
+static const char *apkenv_ldpaths[LDPATH_MAX + 1];
+
+static char apkenv_ldpreloads_buf[LDPRELOAD_BUFSIZE];
+static const char *apkenv_ldpreload_names[LDPRELOAD_MAX + 1];
+
+static soinfo *apkenv_preloads[LDPRELOAD_MAX + 1];
+
+#if LINKER_DEBUG
+int apkenv_debug_verbosity = 5;
+#endif
+
+static int apkenv_pid;
+
+/* This boolean is set if the program being loaded is setuid */
+static int apkenv_program_is_setuid;
+
+#if STATS
+struct _link_stats apkenv_linker_stats;
+#endif
+
+#if COUNT_PAGES
+unsigned apkenv_bitmask[4096];
+#endif
+
+#ifndef PT_ARM_EXIDX
+#define PT_ARM_EXIDX 0x70000001 /* .ARM.exidx segment */
+#endif
+
+static char apkenv_tmp_err_buf[768];
+static char apkenv___linker_dl_err_buf[768];
+#define DL_ERR(fmt, x...) \
+ do { \
+ format_buffer(apkenv___linker_dl_err_buf, sizeof(apkenv___linker_dl_err_buf), \
+ "%s[%d]: " fmt, __func__, __LINE__, ##x); \
+ ERROR(fmt "\n", ##x); \
+ } while(0)
+
+const char *apkenv_linker_get_error(void)
+{
+ return (const char *)&apkenv___linker_dl_err_buf[0];
+}
+
+enum {
+ WRAPPER_DYNHOOK,
+ WRAPPER_THUMB_INJECTION,
+ WRAPPER_UNHOOKED,
+ WRAPPER_ARM_INJECTION,
+};
+
+static void
+trace(const char *const symbol)
+{
+ printf("trace: %s\n", symbol);
+}
+
+__asm__(
+ "wrapper_start: nop\n"
+ "wrapper_symbol: pushl $0xFAFBFCFD\n"
+ "wrapper_trace: movl $0xFAFBFCFD, %eax\ncall *%eax\npop %eax\n"
+ "wrapper_call: movl $0xFAFBFCFD, %eax\njmp *%eax\n"
+ "wrapper_end: nop\n"
+);
+
+extern char wrapper_start, wrapper_symbol, wrapper_trace, wrapper_call, wrapper_end;
+
+void*
+create_wrapper(const char *const symbol, void *function, int wrapper_type)
+{
+ const size_t sz = &wrapper_end - &wrapper_start;
+ unsigned char *fun = mmap(NULL, sz, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
+ assert(fun != MAP_FAILED);
+ memcpy(fun, &wrapper_start, sz);
+ memcpy(fun + (&wrapper_symbol - &wrapper_start) + 1, &symbol, sizeof(symbol));
+ memcpy(fun + (&wrapper_trace - &wrapper_start) + 1, (uintptr_t[]){ (uintptr_t)trace }, sizeof(uintptr_t));
+ memcpy(fun + (&wrapper_call - &wrapper_start) + 1, &function, sizeof(function));
+ mprotect(fun, sz, PROT_READ | PROT_EXEC);
+ return fun;
+}
+
+/*
+ * This function is an empty stub where GDB locates a breakpoint to get notified
+ * about linker activity.
+ */
+//extern void __attribute__((noinline)) rtld_db_dlactivity(void);
+
+//static struct r_debug _r_debug = {1, NULL, &rtld_db_dlactivity,
+// RT_CONSISTENT, 0};
+/* apkenv */
+#define rtld_db_dlactivity() ((void (*)(void))_r_debug.r_brk)()
+
+static struct link_map *apkenv_r_debug_head, *apkenv_r_debug_tail;
+
+static pthread_mutex_t apkenv__r_debug_lock = PTHREAD_MUTEX_INITIALIZER;
+
+static void apkenv_insert_soinfo_into_debug_map(soinfo * info)
+{
+ struct link_map * map;
+
+ /* Copy the necessary fields into the debug structure.
+ */
+ map = &(info->linkmap);
+ map->l_addr = info->base;
+ map->l_name = info->fullpath;
+ map->l_ld = (void *)info->dynamic;
+
+ /* Stick the new library at the end of the list.
+ * gdb tends to care more about libc than it does
+ * about leaf libraries, and ordering it this way
+ * reduces the back-and-forth over the wire.
+ */
+ if (apkenv_r_debug_tail) {
+ apkenv_r_debug_tail->l_next = map;
+ map->l_prev = apkenv_r_debug_tail;
+ map->l_next = 0;
+ } else {
+ /* apkenv: for sending to gdb */
+ apkenv_r_debug_head = map;
+ map->l_prev = 0;
+ map->l_next = 0;
+ }
+ apkenv_r_debug_tail = map;
+}
+
+static void apkenv_remove_soinfo_from_debug_map(soinfo * info)
+{
+ struct link_map * map = &(info->linkmap);
+
+ if (apkenv_r_debug_tail == map)
+ apkenv_r_debug_tail = map->l_prev;
+
+ if (map->l_prev) map->l_prev->l_next = map->l_next;
+ if (map->l_next) map->l_next->l_prev = map->l_prev;
+}
+
+void apkenv_notify_gdb_of_libraries(void);
+
+static void apkenv_notify_gdb_of_load(soinfo * info)
+{
+ if (info->flags & FLAG_EXE) {
+ // GDB already knows about the main executable
+ return;
+ }
+
+ pthread_mutex_lock(&apkenv__r_debug_lock);
+
+ _r_debug.r_state = RT_ADD;
+ rtld_db_dlactivity();
+
+ apkenv_insert_soinfo_into_debug_map(info);
+
+ _r_debug.r_state = RT_CONSISTENT;
+ rtld_db_dlactivity();
+
+ apkenv_notify_gdb_of_libraries();
+ pthread_mutex_unlock(&apkenv__r_debug_lock);
+}
+
+static void apkenv_notify_gdb_of_unload(soinfo * info)
+{
+ if (info->flags & FLAG_EXE) {
+ // GDB already knows about the main executable
+ return;
+ }
+
+ pthread_mutex_lock(&apkenv__r_debug_lock);
+
+ _r_debug.r_state = RT_DELETE;
+ rtld_db_dlactivity();
+
+ apkenv_remove_soinfo_from_debug_map(info);
+
+ _r_debug.r_state = RT_CONSISTENT;
+ rtld_db_dlactivity();
+
+ apkenv_notify_gdb_of_libraries();
+ pthread_mutex_unlock(&apkenv__r_debug_lock);
+}
+
+void apkenv_notify_gdb_of_libraries(void)
+{
+ struct link_map *tmap = _r_debug.r_map;
+ while (tmap->l_next != NULL)
+ tmap = tmap->l_next;
+
+ _r_debug.r_state = RT_ADD;
+ rtld_db_dlactivity();
+
+ /* append android libs before notifying gdb */
+ tmap->l_next = apkenv_r_debug_head;
+ apkenv_r_debug_head->l_prev = tmap;
+
+ _r_debug.r_state = RT_CONSISTENT;
+ rtld_db_dlactivity();
+
+ /* restore so that ld-linux doesn't freak out */
+ tmap->l_next = NULL;
+}
+
+static soinfo *apkenv_alloc_info(const char *name)
+{
+ soinfo *si;
+
+ if(strlen(name) >= SOINFO_NAME_LEN) {
+ DL_ERR("%5d library name %s too long", apkenv_pid, name);
+ return NULL;
+ }
+
+ /* The apkenv_freelist is populated when we call apkenv_free_info(), which in turn is
+ done only by dlclose(), which is not likely to be used.
+ */
+ if (!apkenv_freelist) {
+ if(apkenv_socount == SO_MAX) {
+ DL_ERR("%5d too many libraries when loading %s", apkenv_pid, name);
+ return NULL;
+ }
+ apkenv_freelist = apkenv_sopool + apkenv_socount++;
+ apkenv_freelist->next = NULL;
+ }
+
+ si = apkenv_freelist;
+ apkenv_freelist = apkenv_freelist->next;
+
+ /* Make sure we get a clean block of soinfo */
+ memset(si, 0, sizeof(soinfo));
+ apkenv_strlcpy((char*) si->name, name, sizeof(si->name));
+ apkenv_sonext->next = si;
+ si->next = NULL;
+ si->refcount = 0;
+ apkenv_sonext = si;
+
+ TRACE("%5d name %s: allocated soinfo @ %p\n", apkenv_pid, name, si);
+ return si;
+}
+
+static void apkenv_free_info(soinfo *si)
+{
+ soinfo *prev = NULL, *trav;
+
+ TRACE("%5d name %s: freeing soinfo @ %p\n", apkenv_pid, si->name, si);
+
+ for(trav = apkenv_solist; trav != NULL; trav = trav->next){
+ if (trav == si)
+ break;
+ prev = trav;
+ }
+ if (trav == NULL) {
+ /* si was not ni apkenv_solist */
+ DL_ERR("%5d name %s is not in apkenv_solist!", apkenv_pid, si->name);
+ return;
+ }
+
+ /* prev will never be NULL, because the first entry in apkenv_solist is
+ always the static apkenv_libdl_info.
+ */
+ prev->next = si->next;
+ if (si == apkenv_sonext) apkenv_sonext = prev;
+ si->next = apkenv_freelist;
+ apkenv_freelist = si;
+}
+
+static const char *apkenv_addr_to_name(unsigned addr)
+{
+ soinfo *si;
+
+ for(si = apkenv_solist; si != 0; si = si->next){
+ if((addr >= si->base) && (addr < (si->base + si->size))) {
+ return si->name;
+ }
+ }
+
+ return "";
+}
+
+/* For a given PC, find the .so that it belongs to.
+ * Returns the base address of the .ARM.exidx section
+ * for that .so, and the number of 8-byte entries
+ * in that section (via *pcount).
+ *
+ * Intended to be called by libc's __gnu_Unwind_Find_exidx().
+ *
+ * This function is exposed via dlfcn.c and libdl.so.
+ */
+#ifdef ANDROID_ARM_LINKER
+_Unwind_Ptr apkenv_android_dl_unwind_find_exidx(_Unwind_Ptr pc, int *pcount)
+{
+ soinfo *si;
+ unsigned addr = (unsigned)pc;
+
+ for (si = apkenv_solist; si != 0; si = si->next){
+ if ((addr >= si->base) && (addr < (si->base + si->size))) {
+ *pcount = si->ARM_exidx_count;
+ return (_Unwind_Ptr)(si->base + (unsigned long)si->ARM_exidx);
+ }
+ }
+ *pcount = 0;
+ return NULL;
+}
+#elif defined(ANDROID_X86_LINKER)
+/* Here, we only have to provide a callback to iterate across all the
+ * loaded libraries. gcc_eh does the rest. */
+int
+apkenv_android_dl_iterate_phdr(int (*cb)(struct dl_phdr_info *info, size_t size, void *data),
+ void *data)
+{
+ soinfo *si;
+ struct dl_phdr_info dl_info;
+ int rv = 0;
+
+ for (si = apkenv_solist; si != NULL; si = si->next) {
+ dl_info.dlpi_addr = si->linkmap.l_addr;
+ dl_info.dlpi_name = si->linkmap.l_name;
+ dl_info.dlpi_phdr = (void*)si->phdr;
+ dl_info.dlpi_phnum = si->phnum;
+ rv = cb(&dl_info, sizeof (struct dl_phdr_info), data);
+ if (rv != 0)
+ break;
+ }
+ return rv;
+}
+#endif
+
+static Elf32_Sym *apkenv__elf_lookup(soinfo *si, unsigned hash, const char *name)
+{
+ Elf32_Sym *s;
+ Elf32_Sym *symtab = si->symtab;
+ const char *strtab = si->strtab;
+ unsigned n;
+
+ TRACE_TYPE(LOOKUP, "%5d SEARCH %s in %s@0x%08x %08x %d\n", apkenv_pid,
+ name, si->name, si->base, hash, hash % si->nbucket);
+ if (si->nbucket == 0) {
+ return NULL;
+ }
+ n = hash % si->nbucket;
+
+ for(n = si->bucket[hash % si->nbucket]; n != 0; n = si->chain[n]){
+ s = symtab + n;
+ if(strcmp(strtab + s->st_name, name)) continue;
+
+ /* only concern ourselves with global and weak symbol definitions */
+ switch(ELF32_ST_BIND(s->st_info)){
+ case STB_GLOBAL:
+ case STB_WEAK:
+ /* no section == undefined */
+ if(s->st_shndx == 0) continue;
+
+ TRACE_TYPE(LOOKUP, "%5d FOUND %s in %s (%08x) %d\n", apkenv_pid,
+ name, si->name, s->st_value, s->st_size);
+ return s;
+ }
+ }
+
+ return NULL;
+}
+
+static unsigned apkenv_elfhash(const char *_name)
+{
+ const unsigned char *name = (const unsigned char *) _name;
+ unsigned h = 0, g;
+
+ while(*name) {
+ h = (h << 4) + *name++;
+ g = h & 0xf0000000;
+ h ^= g;
+ h ^= g >> 24;
+ }
+ return h;
+}
+
+const char *apkenv_last_library_used = NULL;
+
+static Elf32_Sym *
+apkenv__do_lookup(soinfo *si, const char *name, unsigned *base)
+{
+ unsigned elf_hash = apkenv_elfhash(name);
+ Elf32_Sym *s;
+ unsigned *d;
+ soinfo *lsi = si;
+ int i;
+
+ /* Look for symbols in the local scope (the object who is
+ * searching). This happens with C++ templates on i386 for some
+ * reason.
+ *
+ * Notes on weak symbols:
+ * The ELF specs are ambigious about treatment of weak definitions in
+ * dynamic linking. Some systems return the first definition found
+ * and some the first non-weak definition. This is system dependent.
+ * Here we return the first definition found for simplicity. */
+
+ s = apkenv__elf_lookup(si, elf_hash, name);
+ if(s != NULL)
+ goto done;
+
+ /* Next, look for it in the apkenv_preloads list */
+ for(i = 0; apkenv_preloads[i] != NULL; i++) {
+ lsi = apkenv_preloads[i];
+ s = apkenv__elf_lookup(lsi, elf_hash, name);
+ if(s != NULL)
+ goto done;
+ }
+
+ for(d = si->dynamic; *d; d += 2) {
+ if(d[0] == DT_NEEDED){
+ lsi = (soinfo *)d[1];
+ if (!apkenv_validate_soinfo(lsi)) {
+ DL_ERR("%5d bad DT_NEEDED pointer in %s",
+ apkenv_pid, si->name);
+ return NULL;
+ }
+
+ DEBUG("%5d %s: looking up %s in %s\n",
+ apkenv_pid, si->name, name, lsi->name);
+ s = apkenv__elf_lookup(lsi, elf_hash, name);
+ if ((s != NULL) && (s->st_shndx != SHN_UNDEF))
+ goto done;
+ }
+ }
+
+#if ALLOW_SYMBOLS_FROM_MAIN
+ /* If we are resolving relocations while dlopen()ing a library, it's OK for
+ * the library to resolve a symbol that's defined in the executable itself,
+ * although this is rare and is generally a bad idea.
+ */
+ if (apkenv_somain) {
+ lsi = apkenv_somain;
+ DEBUG("%5d %s: looking up %s in executable %s\n",
+ apkenv_pid, si->name, name, lsi->name);
+ s = apkenv__elf_lookup(lsi, elf_hash, name);
+ }
+#endif
+
+done:
+ if(s != NULL) {
+ TRACE_TYPE(LOOKUP, "%5d si %s sym %s s->st_value = 0x%08x, "
+ "found in %s, base = 0x%08x\n",
+ apkenv_pid, si->name, name, s->st_value, lsi->name, lsi->base);
+ apkenv_last_library_used = lsi->name;
+ *base = lsi->base;
+ return s;
+ }
+
+ return NULL;
+}
+
+/* This is used by dl_sym(). It performs symbol lookup only within the
+ specified soinfo object and not in any of its dependencies.
+ */
+Elf32_Sym *apkenv_lookup_in_library(soinfo *si, const char *name)
+{
+ return apkenv__elf_lookup(si, apkenv_elfhash(name), name);
+}
+
+/* This is used by dl_sym(). It performs a global symbol lookup.
+ */
+Elf32_Sym *apkenv_lookup(const char *name, soinfo **found, soinfo *start)
+{
+ unsigned elf_hash = apkenv_elfhash(name);
+ Elf32_Sym *s = NULL;
+ soinfo *si;
+
+ if(start == NULL) {
+ start = apkenv_solist;
+ }
+
+ for(si = start; (s == NULL) && (si != NULL); si = si->next)
+ {
+ if(si->flags & FLAG_ERROR)
+ continue;
+ s = apkenv__elf_lookup(si, elf_hash, name);
+ if (s != NULL) {
+ *found = si;
+ break;
+ }
+ }
+
+ if(s != NULL) {
+ TRACE_TYPE(LOOKUP, "%5d %s s->st_value = 0x%08x, "
+ "si->base = 0x%08x\n", apkenv_pid, name, s->st_value, si->base);
+ return s;
+ }
+
+ return NULL;
+}
+
+soinfo *apkenv_find_containing_library(const void *addr)
+{
+ soinfo *si;
+
+ for(si = apkenv_solist; si != NULL; si = si->next)
+ {
+ if((unsigned)addr >= si->base && (unsigned)addr - si->base < si->size) {
+ return si;
+ }
+ }
+
+ return NULL;
+}
+
+Elf32_Sym *apkenv_find_containing_symbol(const void *addr, soinfo *si)
+{
+ unsigned int i;
+ unsigned soaddr = (unsigned)addr - si->base;
+
+ /* Search the library's symbol table for any defined symbol which
+ * contains this address */
+ for(i=0; i<si->nchain; i++) {
+ Elf32_Sym *sym = &si->symtab[i];
+
+ if(sym->st_shndx != SHN_UNDEF &&
+ soaddr >= sym->st_value &&
+ soaddr < sym->st_value + sym->st_size) {
+ return sym;
+ }
+ }
+
+ return NULL;
+}
+
+#if 0
+static void dump(soinfo *si)
+{
+ Elf32_Sym *s = si->symtab;
+ unsigned n;
+
+ for(n = 0; n < si->nchain; n++) {
+ TRACE("%5d %04d> %08x: %02x %04x %08x %08x %s\n", apkenv_pid, n, s,
+ s->st_info, s->st_shndx, s->st_value, s->st_size,
+ si->strtab + s->st_name);
+ s++;
+ }
+}
+#endif
+
+#define SOPATH_MAX 8
+
+static const char *apkenv_sopaths[SOPATH_MAX + 1] = {
+ "/vendor/lib",
+ "/system/lib",
+ // APKENV_PREFIX "/lib/" APKENV_TARGET "/bionic/",
+#ifdef APKENV_LOCAL_BIONIC_PATH
+ APKENV_LOCAL_BIONIC_PATH,
+#endif
+};
+
+int apkenv_add_sopath(const char *path)
+{
+ int i;
+ for (i = 0; i < SOPATH_MAX; i++) {
+ if (apkenv_sopaths[i] == NULL) {
+ apkenv_sopaths[i] = path;
+ return 0;
+ }
+
+ if (strcmp(path, apkenv_sopaths[i]) == 0)
+ return 0;
+ }
+
+ ERROR("too many apkenv_sopaths\n");
+ return -1;
+}
+
+static int apkenv__open_lib(const char *name)
+{
+ int fd;
+ struct stat filestat;
+
+ if ((stat(name, &filestat) >= 0) && S_ISREG(filestat.st_mode)) {
+ if ((fd = open(name, O_RDONLY)) >= 0)
+ return fd;
+ }
+
+ return -1;
+}
+
+static int apkenv_open_library(const char *name, char *fullpath)
+{
+ int fd;
+ char buf[512];
+ const char **path;
+ int n;
+
+ TRACE("[ %5d opening %s ]\n", apkenv_pid, name);
+
+ if(name == 0) return -1;
+ if(strlen(name) > 256) return -1;
+
+ strcpy(fullpath, name);
+
+ if ((name[0] == '/') && ((fd = apkenv__open_lib(name)) >= 0))
+ return fd;
+
+ if (((fd = apkenv__open_lib(name)) >= 0))
+ return fd;
+
+ for (path = apkenv_ldpaths; *path; path++) {
+ n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
+ if (n < 0 || n >= (int)sizeof(buf)) {
+ WARN("Ignoring very long library path: %s/%s\n", *path, name);
+ continue;
+ }
+ if ((fd = apkenv__open_lib(buf)) >= 0) {
+ strcpy(fullpath, buf);
+ return fd;
+ }
+ }
+ for (path = apkenv_sopaths; *path; path++) {
+ n = format_buffer(buf, sizeof(buf), "%s/%s", *path, name);
+ if (n < 0 || n >= (int)sizeof(buf)) {
+ WARN("Ignoring very long library path: %s/%s\n", *path, name);
+ continue;
+ }
+ if ((fd = apkenv__open_lib(buf)) >= 0) {
+ strcpy(fullpath, buf);
+ return fd;
+ }
+ }
+
+ return -1;
+}
+
+typedef struct {
+ long mmap_addr;
+ char tag[4]; /* 'P', 'R', 'E', ' ' */
+} prelink_info_t;
+
+/* Returns the requested base address if the library is prelinked,
+ * and 0 otherwise. */
+static unsigned long
+apkenv_is_prelinked(int fd, const char *name)
+{
+ off_t sz;
+ prelink_info_t info;
+
+ sz = lseek(fd, -sizeof(prelink_info_t), SEEK_END);
+ if (sz < 0) {
+ DL_ERR("lseek() failed!");
+ return 0;
+ }
+
+ if (read(fd, &info, sizeof(info)) != sizeof(info)) {
+ WARN("Could not read prelink_info_t structure for `%s`\n", name);
+ return 0;
+ }
+
+ if (strncmp(info.tag, "PRE ", 4)) {
+ WARN("`%s` is not a prelinked library\n", name);
+ return 0;
+ }
+
+ return (unsigned long)info.mmap_addr;
+}
+
+/* apkenv_verify_elf_object
+ * Verifies if the object @ base is a valid ELF object
+ *
+ * Args:
+ *
+ * Returns:
+ * 0 on success
+ * -1 if no valid ELF object is found @ base.
+ */
+static int
+apkenv_verify_elf_object(void *base, const char *name)
+{
+ Elf32_Ehdr *hdr = (Elf32_Ehdr *) base;
+
+ if (hdr->e_ident[EI_MAG0] != ELFMAG0) return -1;
+ if (hdr->e_ident[EI_MAG1] != ELFMAG1) return -1;
+ if (hdr->e_ident[EI_MAG2] != ELFMAG2) return -1;
+ if (hdr->e_ident[EI_MAG3] != ELFMAG3) return -1;
+
+ /* TODO: Should we verify anything else in the header? */
+#ifdef ANDROID_ARM_LINKER
+ if (hdr->e_machine != EM_ARM) return -1;
+#elif defined(ANDROID_X86_LINKER)
+ if (hdr->e_machine != EM_386) return -1;
+#endif
+ return 0;
+}
+
+
+/* apkenv_get_lib_extents
+ * Retrieves the base (*base) address where the ELF object should be
+ * mapped and its overall memory size (*total_sz).
+ *
+ * Args:
+ * fd: Opened file descriptor for the library
+ * name: The name of the library
+ * _hdr: Pointer to the header page of the library
+ * total_sz: Total size of the memory that should be allocated for
+ * this library
+ *
+ * Returns:
+ * -1 if there was an error while trying to get the lib extents.
+ * The possible reasons are:
+ * - Could not determine if the library was prelinked.
+ * - The library provided is not a valid ELF object
+ * 0 if the library did not request a specific base offset (normal
+ * for non-prelinked libs)
+ * > 0 if the library requests a specific address to be mapped to.
+ * This indicates a pre-linked library.
+ */
+static unsigned
+apkenv_get_lib_extents(int fd, const char *name, void *__hdr, unsigned *total_sz)
+{
+ unsigned req_base;
+ unsigned min_vaddr = 0xffffffff;
+ unsigned max_vaddr = 0;
+ unsigned char *_hdr = (unsigned char *)__hdr;
+ Elf32_Ehdr *ehdr = (Elf32_Ehdr *)_hdr;
+ Elf32_Phdr *phdr;
+ int cnt;
+
+ TRACE("[ %5d Computing extents for '%s'. ]\n", apkenv_pid, name);
+ if (apkenv_verify_elf_object(_hdr, name) < 0) {
+ DL_ERR("%5d - %s is not a valid ELF object", apkenv_pid, name);
+ return (unsigned)-1;
+ }
+
+ req_base = (unsigned) apkenv_is_prelinked(fd, name);
+ if (req_base == (unsigned)-1)
+ return -1;
+ else if (req_base != 0) {
+ TRACE("[ %5d - Prelinked library '%s' requesting base @ 0x%08x ]\n",
+ apkenv_pid, name, req_base);
+ } else {
+ TRACE("[ %5d - Non-prelinked library '%s' found. ]\n", apkenv_pid, name);
+ }
+
+ phdr = (Elf32_Phdr *)(_hdr + ehdr->e_phoff);
+
+ /* find the min/max p_vaddrs from all the PT_LOAD segments so we can
+ * get the range. */
+ for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
+ if (phdr->p_type == PT_LOAD) {
+ if ((phdr->p_vaddr + phdr->p_memsz) > max_vaddr)
+ max_vaddr = phdr->p_vaddr + phdr->p_memsz;
+ if (phdr->p_vaddr < min_vaddr)
+ min_vaddr = phdr->p_vaddr;
+ }
+ }
+
+ if ((min_vaddr == 0xffffffff) && (max_vaddr == 0)) {
+ DL_ERR("%5d - No loadable segments found in %s.", apkenv_pid, name);
+ return (unsigned)-1;
+ }
+
+ /* truncate min_vaddr down to page boundary */
+ min_vaddr &= ~PAGE_MASK;
+
+ /* round max_vaddr up to the next page */
+ max_vaddr = (max_vaddr + PAGE_SIZE - 1) & ~PAGE_MASK;
+
+ *total_sz = (max_vaddr - min_vaddr);
+ return (unsigned)req_base;
+}
+
+/* apkenv_reserve_mem_region
+ *
+ * This function reserves a chunk of memory to be used for mapping in
+ * a prelinked shared library. We reserve the entire memory region here, and
+ * then the rest of the linker will relocate the individual loadable
+ * segments into the correct locations within this memory range.
+ *
+ * Args:
+ * si->base: The requested base of the allocation.
+ * si->size: The size of the allocation.
+ *
+ * Returns:
+ * -1 on failure, and 0 on success. On success, si->base will contain
+ * the virtual address at which the library will be mapped.
+ */
+
+static int apkenv_reserve_mem_region(soinfo *si)
+{
+ void *base = mmap((void *)si->base, si->size, PROT_NONE,
+ MAP_FIXED | MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
+ if (base == MAP_FAILED) {
+ DL_ERR("%5d can NOT map (%sprelinked) library '%s' at 0x%08x "
+ "as requested, will try general pool: %d (%s)",
+ apkenv_pid, (si->base ? "" : "non-"), si->name, si->base,
+ errno, strerror(errno));
+ return -1;
+ } else if (base != (void *)si->base) {
+ DL_ERR("OOPS: %5d %sprelinked library '%s' mapped at 0x%08x, "
+ "not at 0x%08x", apkenv_pid, (si->base ? "" : "non-"),
+ si->name, (unsigned)base, si->base);
+ munmap(base, si->size);
+ return -1;
+ }
+ return 0;
+}
+
+static int apkenv_alloc_mem_region(soinfo *si)
+{
+ if (si->base) {
+ /* Attempt to mmap a prelinked library. */
+ return apkenv_reserve_mem_region(si);
+ }
+
+ /* This is not a prelinked library, so we use the kernel's default
+ allocator.
+ */
+
+ void *base = mmap(NULL, si->size, PROT_NONE,
+ MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
+ if (base == MAP_FAILED) {
+ DL_ERR("%5d mmap of library '%s' failed: %d (%s)\n",
+ apkenv_pid, si->name,
+ errno, strerror(errno));
+ goto err;
+ }
+ si->base = (unsigned) base;
+ PRINT("%5d mapped library '%s' to %08x via kernel allocator.\n",
+ apkenv_pid, si->name, si->base);
+ return 0;
+
+err:
+ DL_ERR("OOPS: %5d cannot map library '%s'. no vspace available.",
+ apkenv_pid, si->name);
+ return -1;
+}
+
+#define MAYBE_MAP_FLAG(x,from,to) (((x) & (from)) ? (to) : 0)
+#define PFLAGS_TO_PROT(x) (MAYBE_MAP_FLAG((x), PF_X, PROT_EXEC) | \
+ MAYBE_MAP_FLAG((x), PF_R, PROT_READ) | \
+ MAYBE_MAP_FLAG((x), PF_W, PROT_WRITE))
+/* apkenv_load_segments
+ *
+ * This function loads all the loadable (PT_LOAD) segments into memory
+ * at their appropriate memory offsets off the base address.
+ *
+ * Args:
+ * fd: Open file descriptor to the library to load.
+ * header: Pointer to a header page that contains the ELF header.
+ * This is needed since we haven't mapped in the real file yet.
+ * si: ptr to soinfo struct describing the shared object.
+ *
+ * Returns:
+ * 0 on success, -1 on failure.
+ */
+static int
+apkenv_load_segments(int fd, void *header, soinfo *si)
+{
+ Elf32_Ehdr *ehdr = (Elf32_Ehdr *)header;
+ Elf32_Phdr *phdr = (Elf32_Phdr *)((unsigned char *)header + ehdr->e_phoff);
+ Elf32_Addr base = (Elf32_Addr) si->base;
+ int cnt;
+ unsigned len;
+ Elf32_Addr tmp;
+ unsigned char *pbase;
+ unsigned char *extra_base;
+ unsigned extra_len;
+ unsigned total_sz = 0;
+
+ si->wrprotect_start = 0xffffffff;
+ si->wrprotect_end = 0;
+
+ TRACE("[ %5d - Begin loading segments for '%s' @ 0x%08x ]\n",
+ apkenv_pid, si->name, (unsigned)si->base);
+ /* Now go through all the PT_LOAD segments and map them into memory
+ * at the appropriate locations. */
+ for (cnt = 0; cnt < ehdr->e_phnum; ++cnt, ++phdr) {
+ if (phdr->p_type == PT_LOAD) {
+ DEBUG_DUMP_PHDR(phdr, "PT_LOAD", apkenv_pid);
+ /* we want to map in the segment on a page boundary */
+ tmp = base + (phdr->p_vaddr & (~PAGE_MASK));
+ /* add the # of bytes we masked off above to the total length. */
+ len = phdr->p_filesz + (phdr->p_vaddr & PAGE_MASK);
+
+ TRACE("[ %d - Trying to load segment from '%s' @ 0x%08x "
+ "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x ]\n", apkenv_pid, si->name,
+ (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
+ pbase = mmap((void *)tmp, len, PFLAGS_TO_PROT(phdr->p_flags),
+ MAP_PRIVATE | MAP_FIXED, fd,
+ phdr->p_offset & (~PAGE_MASK));
+ if (pbase == MAP_FAILED) {
+ DL_ERR("%d failed to map segment from '%s' @ 0x%08x (0x%08x). "
+ "p_vaddr=0x%08x p_offset=0x%08x", apkenv_pid, si->name,
+ (unsigned)tmp, len, phdr->p_vaddr, phdr->p_offset);
+ goto fail;
+ }
+
+ /* If 'len' didn't end on page boundary, and it's a writable
+ * segment, zero-fill the rest. */
+ if ((len & PAGE_MASK) && (phdr->p_flags & PF_W))
+ memset((void *)(pbase + len), 0, PAGE_SIZE - (len & PAGE_MASK));
+
+ /* Check to see if we need to extend the map for this segment to
+ * cover the diff between filesz and memsz (i.e. for bss).
+ *
+ * base _+---------------------+ page boundary
+ * . .
+ * | |
+ * . .
+ * pbase _+---------------------+ page boundary
+ * | |
+ * . .
+ * base + p_vaddr _| |
+ * . \ \ .
+ * . | filesz | .
+ * pbase + len _| / | |
+ * <0 pad> . . .
+ * extra_base _+------------|--------+ page boundary
+ * / . . .
+ * | . . .
+ * | +------------|--------+ page boundary
+ * extra_len-> | | | |
+ * | . | memsz .
+ * | . | .
+ * \ _| / |
+ * . .
+ * | |
+ * _+---------------------+ page boundary
+ */
+ tmp = (Elf32_Addr)(((unsigned)pbase + len + PAGE_SIZE - 1) &
+ (~PAGE_MASK));
+ if (tmp < (base + phdr->p_vaddr + phdr->p_memsz)) {
+ extra_len = base + phdr->p_vaddr + phdr->p_memsz - tmp;
+ TRACE("[ %5d - Need to extend segment from '%s' @ 0x%08x "
+ "(0x%08x) ]\n", apkenv_pid, si->name, (unsigned)tmp, extra_len);
+ /* map in the extra page(s) as anonymous into the range.
+ * This is probably not necessary as we already mapped in
+ * the entire region previously, but we just want to be
+ * sure. This will also set the right flags on the region
+ * (though we can probably accomplish the same thing with
+ * mprotect).
+ */
+ extra_base = mmap((void *)tmp, extra_len,
+ PFLAGS_TO_PROT(phdr->p_flags),
+ MAP_PRIVATE | MAP_FIXED | MAP_ANONYMOUS,
+ -1, 0);
+ if (extra_base == MAP_FAILED) {
+ DL_ERR("[ %5d - failed to extend segment from '%s' @ 0x%08x"
+ " (0x%08x) ]", apkenv_pid, si->name, (unsigned)tmp,
+ extra_len);
+ goto fail;
+ }
+ /* TODO: Check if we need to memset-0 this region.
+ * Anonymous mappings are zero-filled copy-on-writes, so we
+ * shouldn't need to. */
+ TRACE("[ %5d - Segment from '%s' extended @ 0x%08x "
+ "(0x%08x)\n", apkenv_pid, si->name, (unsigned)extra_base,
+ extra_len);
+ }
+ /* set the len here to show the full extent of the segment we
+ * just loaded, mostly for debugging */
+ len = (((unsigned)base + phdr->p_vaddr + phdr->p_memsz +
+ PAGE_SIZE - 1) & (~PAGE_MASK)) - (unsigned)pbase;
+ TRACE("[ %5d - Successfully loaded segment from '%s' @ 0x%08x "
+ "(0x%08x). p_vaddr=0x%08x p_offset=0x%08x\n", apkenv_pid, si->name,
+ (unsigned)pbase, len, phdr->p_vaddr, phdr->p_offset);
+ total_sz += len;
+ /* Make the section writable just in case we'll have to write to
+ * it during relocation (i.e. text segment). However, we will
+ * remember what range of addresses should be write protected.
+ *
+ */
+ if (!(phdr->p_flags & PF_W)) {
+ if ((unsigned)pbase < si->wrprotect_start)
+ si->wrprotect_start = (unsigned)pbase;
+ if (((unsigned)pbase + len) > si->wrprotect_end)
+ si->wrprotect_end = (unsigned)pbase + len;
+ mprotect(pbase, len,
+ PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
+ }
+ } else if (phdr->p_type == PT_DYNAMIC) {
+ DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", apkenv_pid);
+ /* this segment contains the dynamic linking information */
+ si->dynamic = (unsigned *)(base + phdr->p_vaddr);
+ } else if (phdr->p_type == PT_GNU_RELRO) {
+ if ((phdr->p_vaddr >= si->size)
+ || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
+ || ((base + phdr->p_vaddr + phdr->p_memsz) < base)) {
+ DL_ERR("%d invalid GNU_RELRO in '%s' "
+ "p_vaddr=0x%08x p_memsz=0x%08x", apkenv_pid, si->name,
+ phdr->p_vaddr, phdr->p_memsz);
+ goto fail;
+ }
+ si->gnu_relro_start = (Elf32_Addr) (base + phdr->p_vaddr);
+ si->gnu_relro_len = (unsigned) phdr->p_memsz;
+ } else {
+#ifdef ANDROID_ARM_LINKER
+ if (phdr->p_type == PT_ARM_EXIDX) {
+ DEBUG_DUMP_PHDR(phdr, "PT_ARM_EXIDX", apkenv_pid);
+ /* exidx entries (used for stack unwinding) are 8 bytes each.
+ */
+ si->ARM_exidx = (unsigned *)phdr->p_vaddr;
+ si->ARM_exidx_count = phdr->p_memsz / 8;
+ }
+#endif
+ }
+
+ }
+
+ /* Sanity check */
+ if (total_sz > si->size) {
+ DL_ERR("%5d - Total length (0x%08x) of mapped segments from '%s' is "
+ "greater than what was allocated (0x%08x). THIS IS BAD!",
+ apkenv_pid, total_sz, si->name, si->size);
+ goto fail;
+ }
+
+ TRACE("[ %5d - Finish loading segments for '%s' @ 0x%08x. "
+ "Total memory footprint: 0x%08x bytes ]\n", apkenv_pid, si->name,
+ (unsigned)si->base, si->size);
+ return 0;
+
+fail:
+ /* We can just blindly unmap the entire region even though some things
+ * were mapped in originally with anonymous and others could have been
+ * been mapped in from the file before we failed. The kernel will unmap
+ * all the pages in the range, irrespective of how they got there.
+ */
+ munmap((void *)si->base, si->size);
+ si->flags |= FLAG_ERROR;
+ return -1;
+}
+
+/* TODO: Implement this to take care of the fact that Android ARM
+ * ELF objects shove everything into a single loadable segment that has the
+ * write bit set. wr_offset is then used to set non-(data|bss) pages to be
+ * non-writable.
+ */
+#if 0
+static unsigned
+get_wr_offset(int fd, const char *name, Elf32_Ehdr *ehdr)
+{
+ Elf32_Shdr *shdr_start;
+ Elf32_Shdr *shdr;
+ int shdr_sz = ehdr->e_shnum * sizeof(Elf32_Shdr);
+ int cnt;
+ unsigned wr_offset = 0xffffffff;
+
+ shdr_start = mmap(0, shdr_sz, PROT_READ, MAP_PRIVATE, fd,
+ ehdr->e_shoff & (~PAGE_MASK));
+ if (shdr_start == MAP_FAILED) {
+ WARN("%5d - Could not read section header info from '%s'. Will not "
+ "not be able to determine write-protect offset.\n", apkenv_pid, name);
+ return (unsigned)-1;
+ }
+
+ for(cnt = 0, shdr = shdr_start; cnt < ehdr->e_shnum; ++cnt, ++shdr) {
+ if ((shdr->sh_type != SHT_NULL) && (shdr->sh_flags & SHF_WRITE) &&
+ (shdr->sh_addr < wr_offset)) {
+ wr_offset = shdr->sh_addr;
+ }
+ }
+
+ munmap(shdr_start, shdr_sz);
+ return wr_offset;
+}
+#endif
+
+static soinfo *
+apkenv_load_library(const char *name)
+{
+ char fullpath[512];
+ int fd = apkenv_open_library(name, fullpath);
+ int cnt;
+ unsigned ext_sz;
+ unsigned req_base;
+ const char *bname;
+ soinfo *si = NULL;
+ Elf32_Ehdr *hdr;
+
+ if(fd == -1) {
+#if !LINKER_DEBUG
+ if (!is_lib_optional(name))
+#endif
+ DL_ERR("Library '%s' not found", name);
+ return NULL;
+ }
+
+ off_t end;
+ if ((end = lseek(fd, 0, SEEK_END)) < 0) {
+ DL_ERR("lseek() failed!");
+ goto fail;
+ }
+
+ uint8_t *bytes;
+ if (!(bytes = calloc(1, end))) {
+ DL_ERR("calloc() failed!");
+ goto fail;
+ }
+
+ /* We have to read the ELF header to figure out what to do with this image
+ */
+ if (lseek(fd, 0, SEEK_SET) < 0) {
+ DL_ERR("lseek() failed!");
+ goto fail;
+ }
+
+
+ if ((cnt = read(fd, bytes, end)) < 0) {
+ DL_ERR("read() failed!");
+ goto fail;
+ }
+
+ /* Parse the ELF header and get the size of the memory footprint for
+ * the library */
+ req_base = apkenv_get_lib_extents(fd, name, bytes, &ext_sz);
+ if (req_base == (unsigned)-1)
+ goto fail;
+ TRACE("[ %5d - '%s' (%s) wants base=0x%08x sz=0x%08x ]\n", apkenv_pid, name,
+ (req_base ? "prelinked" : "not pre-linked"), req_base, ext_sz);
+
+ /* Now configure the soinfo struct where we'll store all of our data
+ * for the ELF object. If the loading fails, we waste the entry, but
+ * same thing would happen if we failed during linking. Configuring the
+ * soinfo struct here is a lot more convenient.
+ */
+ bname = strrchr(name, '/');
+ si = apkenv_alloc_info(bname ? bname + 1 : name);
+ if (si == NULL)
+ goto fail;
+
+ /* apkenv */
+ strncpy(si->fullpath, fullpath, sizeof(si->fullpath));
+ si->fullpath[sizeof(si->fullpath) - 1] = 0;
+
+ /* Carve out a chunk of memory where we will map in the individual
+ * segments */
+ si->base = req_base;
+ si->size = ext_sz;
+ si->flags = 0;
+ si->entry = 0;
+ si->dynamic = (unsigned *)-1;
+ if (apkenv_alloc_mem_region(si) < 0)
+ goto fail;
+
+ TRACE("[ %5d allocated memory for %s @ %p (0x%08x) ]\n",
+ apkenv_pid, name, (void *)si->base, (unsigned) ext_sz);
+
+ /* Now actually load the library's segments into right places in memory */
+ if (apkenv_load_segments(fd, bytes, si) < 0) {
+ goto fail;
+ }
+
+ /* this might not be right. Technically, we don't even need this info
+ * once we go through 'apkenv_load_segments'. */
+ hdr = (Elf32_Ehdr *)si->base;
+ si->phdr = (Elf32_Phdr *)((unsigned char *)si->base + hdr->e_phoff);
+ si->phnum = hdr->e_phnum;
+ /**/
+
+ close(fd);
+ return si;
+
+fail:
+ if (si) apkenv_free_info(si);
+ close(fd);
+ return NULL;
+}
+
+static soinfo *
+apkenv_init_library(soinfo *si)
+{
+ unsigned wr_offset = 0xffffffff;
+
+ /* At this point we know that whatever is loaded @ base is a valid ELF
+ * shared library whose segments are properly mapped in. */
+ TRACE("[ %5d apkenv_init_library base=0x%08x sz=0x%08x name='%s') ]\n",
+ apkenv_pid, si->base, si->size, si->name);
+ if(apkenv_link_image(si, wr_offset)) {
+ /* We failed to link. However, we can only restore libbase
+ ** if no additional libraries have moved it since we updated it.
+ */
+ munmap((void *)si->base, si->size);
+ return NULL;
+ }
+
+ return si;
+}
+
+soinfo *apkenv_find_library(const char *name)
+{
+ soinfo *si;
+ const char *bname;
+
+#if ALLOW_SYMBOLS_FROM_MAIN
+ if (name == NULL)
+ return apkenv_somain;
+#else
+ if (name == NULL)
+ return NULL;
+#endif
+
+ bname = strrchr(name, '/');
+ bname = bname ? bname + 1 : name;
+
+ for(si = apkenv_solist; si != 0; si = si->next){
+ if(!strcmp(bname, si->name)) {
+ if(si->flags & FLAG_ERROR) {
+ DL_ERR("%5d '%s' failed to load previously", apkenv_pid, bname);
+ return NULL;
+ }
+ if(si->flags & FLAG_LINKED) return si;
+ DL_ERR("OOPS: %5d recursive link to '%s'", apkenv_pid, si->name);
+ return NULL;
+ }
+ }
+
+ TRACE("[ %5d '%s' has not been loaded yet. Locating...]\n", apkenv_pid, name);
+ si = apkenv_load_library(name);
+ if(si == NULL)
+ return NULL;
+ return apkenv_init_library(si);
+}
+
+/* TODO:
+ * notify gdb of unload
+ * for non-prelinked libraries, find a way to decrement libbase
+ */
+static void apkenv_call_destructors(soinfo *si);
+unsigned apkenv_unload_library(soinfo *si)
+{
+ unsigned *d;
+ if (si->refcount == 1) {
+ TRACE("%5d unloading '%s'\n", apkenv_pid, si->name);
+ apkenv_call_destructors(si);
+
+ /*
+ * Make sure that we undo the PT_GNU_RELRO protections we added
+ * in apkenv_link_image. This is needed to undo the DT_NEEDED hack below.
+ */
+ if ((si->gnu_relro_start != 0) && (si->gnu_relro_len != 0)) {
+ Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
+ unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
+ if (mprotect((void *) start, len, PROT_READ | PROT_WRITE) < 0)
+ DL_ERR("%5d %s: could not undo GNU_RELRO protections. "
+ "Expect a crash soon. errno=%d (%s)",
+ apkenv_pid, si->name, errno, strerror(errno));
+
+ }
+
+ for(d = si->dynamic; *d; d += 2) {
+ if(d[0] == DT_NEEDED){
+ soinfo *lsi = (soinfo *)d[1];
+
+ // The next line will segfault if the we don't undo the
+ // PT_GNU_RELRO protections (see comments above and in
+ // apkenv_link_image().
+ d[1] = 0;
+
+ if (apkenv_validate_soinfo(lsi)) {
+ TRACE("%5d %s needs to unload %s\n", apkenv_pid,
+ si->name, lsi->name);
+ apkenv_unload_library(lsi);
+ }
+ else
+ DL_ERR("%5d %s: could not unload dependent library",
+ apkenv_pid, si->name);
+ }
+ }
+
+ munmap((char *)si->base, si->size);
+ apkenv_notify_gdb_of_unload(si);
+ apkenv_free_info(si);
+ si->refcount = 0;
+ }
+ else {
+ si->refcount--;
+ PRINT("%5d not unloading '%s', decrementing refcount to %d\n",
+ apkenv_pid, si->name, si->refcount);
+ }
+ return si->refcount;
+}
+
+#define MIN(a,b) (((a)<(b))?(a):(b))
+
+/* TODO: don't use unsigned for addrs below. It works, but is not
+ * ideal. They should probably be either uint32_t, Elf32_Addr, or unsigned
+ * long.
+ */
+static int apkenv_reloc_library(soinfo *si, Elf32_Rel *rel, unsigned count)
+{
+ Elf32_Sym *symtab = si->symtab;
+ const char *strtab = si->strtab;
+ Elf32_Sym *s;
+ unsigned base;
+ Elf32_Rel *start = rel;
+ unsigned idx;
+
+ for (idx = 0; idx < count; ++idx) {
+ unsigned type = ELF32_R_TYPE(rel->r_info);
+ unsigned sym = ELF32_R_SYM(rel->r_info);
+ unsigned reloc = (unsigned)(rel->r_offset + si->base);
+ unsigned sym_addr = 0;
+ char *sym_name = NULL;
+ char wrap_sym_name[1024] = { 'b', 'i', 'o', 'n', 'i', 'c', '_' };
+
+ DEBUG("%5d Processing '%s' relocation at index %d\n", apkenv_pid,
+ si->name, idx);
+ if(sym != 0) {
+ sym_name = (char *)(strtab + symtab[sym].st_name);
+ memcpy(wrap_sym_name + 7, sym_name, MIN(sizeof(wrap_sym_name) - 7, strlen(sym_name)));
+ sym_addr = 0;
+ if ((sym_addr = dlsym(RTLD_DEFAULT, wrap_sym_name))) {
+ LINKER_DEBUG_PRINTF("%s hooked symbol %s to %x\n", si->name, wrap_sym_name, sym_addr);
+ } else if ((sym_addr = dlsym(RTLD_DEFAULT, sym_name))) {
+ LINKER_DEBUG_PRINTF("%s hooked symbol %s to %x\n", si->name, sym_name, sym_addr);
+ }
+#if 0
+ if ((sym_addr = (unsigned)apkenv_get_hooked_symbol(sym_name, 1)) != 0) {
+ LINKER_DEBUG_PRINTF("%s hooked symbol %s to %x\n", si->name, sym_name, sym_addr);
+ }
+#endif
+ else
+ {
+ s = apkenv__do_lookup(si, sym_name, &base);
+ }
+ if(sym_addr != 0)
+ {
+ sym_addr = (unsigned)create_wrapper(sym_name, (void*)sym_addr, WRAPPER_UNHOOKED);
+ } else
+ if(s == NULL) {
+ /* We only allow an undefined symbol if this is a weak
+ reference.. */
+ s = &symtab[sym];
+ if (ELF32_ST_BIND(s->st_info) != STB_WEAK) {
+ DL_ERR("%5d cannot locate '%s'...\n", apkenv_pid, sym_name);
+ return -1;
+ }
+
+ /* IHI0044C AAELF 4.5.1.1:
+
+ Libraries are not searched to resolve weak references.
+ It is not an error for a weak reference to remain
+ unsatisfied.
+
+ During linking, the value of an undefined weak reference is:
+ - Zero if the relocation type is absolute
+ - The address of the place if the relocation is pc-relative
+ - The address of nominial base address if the relocation
+ type is base-relative.
+ */
+
+ switch (type) {
+#if defined(ANDROID_ARM_LINKER)
+ case R_ARM_JUMP_SLOT:
+ case R_ARM_GLOB_DAT:
+ case R_ARM_ABS32:
+ case R_ARM_RELATIVE: /* Don't care. */
+ case R_ARM_NONE: /* Don't care. */
+#elif defined(ANDROID_X86_LINKER)
+ case R_386_JUMP_SLOT:
+ case R_386_GLOB_DAT:
+ case R_386_32:
+ case R_386_RELATIVE: /* Dont' care. */
+#endif /* ANDROID_*_LINKER */
+ /* sym_addr was initialized to be zero above or relocation
+ code below does not care about value of sym_addr.
+ No need to do anything. */
+ break;
+
+#if defined(ANDROID_X86_LINKER)
+ case R_386_PC32:
+ sym_addr = reloc;
+ break;
+#endif /* ANDROID_X86_LINKER */
+
+#if defined(ANDROID_ARM_LINKER)
+ case R_ARM_COPY:
+ /* Fall through. Can't really copy if weak symbol is
+ not found in run-time. */
+#endif /* ANDROID_ARM_LINKER */
+ default:
+ DL_ERR("%5d unknown weak reloc type %d @ %p (%d)\n",
+ apkenv_pid, type, rel, (int) (rel - start));
+ return -1;
+ }
+ } else {
+ /* We got a definition. */
+#if 0
+ if((base == 0) && (si->base != 0)){
+ /* linking from libraries to main image is bad */
+ DL_ERR("%5d cannot locate '%s'...",
+ apkenv_pid, strtab + symtab[sym].st_name);
+ return -1;
+ }
+#endif
+ sym_addr = (unsigned)(s->st_value + base);
+ LINKER_DEBUG_PRINTF("%s symbol (from %s) %s to %x\n", si->name, apkenv_last_library_used, sym_name, sym_addr);
+ if(ELF32_ST_TYPE(s->st_info) == STT_FUNC) {
+ sym_addr = (unsigned)create_wrapper(sym_name, (void*)sym_addr, WRAPPER_UNHOOKED);
+ }
+
+ }
+ COUNT_RELOC(RELOC_SYMBOL);
+ } else {
+ s = NULL;
+ }
+
+/* TODO: This is ugly. Split up the relocations by arch into
+ * different files.
+ */
+ switch(type){
+#if defined(ANDROID_ARM_LINKER)
+ case R_ARM_JUMP_SLOT:
+ COUNT_RELOC(RELOC_ABSOLUTE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned*)reloc) = sym_addr;
+ break;
+ case R_ARM_GLOB_DAT:
+ COUNT_RELOC(RELOC_ABSOLUTE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned*)reloc) = sym_addr;
+ break;
+ case R_ARM_ABS32:
+ COUNT_RELOC(RELOC_ABSOLUTE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO ABS %08x <- %08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned*)reloc) += sym_addr;
+ break;
+ case R_ARM_REL32:
+ COUNT_RELOC(RELOC_RELATIVE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO REL32 %08x <- %08x - %08x %s\n", apkenv_pid,
+ reloc, sym_addr, rel->r_offset, sym_name);
+ *((unsigned*)reloc) += sym_addr - rel->r_offset;
+ break;
+#elif defined(ANDROID_X86_LINKER)
+ case R_386_JUMP_SLOT:
+ COUNT_RELOC(RELOC_ABSOLUTE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO JMP_SLOT %08x <- %08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned*)reloc) = sym_addr;
+ break;
+ case R_386_GLOB_DAT:
+ COUNT_RELOC(RELOC_ABSOLUTE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO GLOB_DAT %08x <- %08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned*)reloc) = sym_addr;
+ break;
+#endif /* ANDROID_*_LINKER */
+
+#if defined(ANDROID_ARM_LINKER)
+ case R_ARM_RELATIVE:
+#elif defined(ANDROID_X86_LINKER)
+ case R_386_RELATIVE:
+#endif /* ANDROID_*_LINKER */
+ COUNT_RELOC(RELOC_RELATIVE);
+ MARK(rel->r_offset);
+ if(sym){
+ DL_ERR("%5d odd RELATIVE form...", apkenv_pid);
+ return -1;
+ }
+ TRACE_TYPE(RELO, "%5d RELO RELATIVE %08x <- +%08x\n", apkenv_pid,
+ reloc, si->base);
+ *((unsigned*)reloc) += si->base;
+ break;
+
+#if defined(ANDROID_X86_LINKER)
+ case R_386_32:
+ COUNT_RELOC(RELOC_RELATIVE);
+ MARK(rel->r_offset);
+
+ TRACE_TYPE(RELO, "%5d RELO R_386_32 %08x <- +%08x %s\n", apkenv_pid,
+ reloc, sym_addr, sym_name);
+ *((unsigned *)reloc) += (unsigned)sym_addr;
+ break;
+
+ case R_386_PC32:
+ COUNT_RELOC(RELOC_RELATIVE);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO R_386_PC32 %08x <- "
+ "+%08x (%08x - %08x) %s\n", apkenv_pid, reloc,
+ (sym_addr - reloc), sym_addr, reloc, sym_name);
+ *((unsigned *)reloc) += (unsigned)(sym_addr - reloc);
+ break;
+#endif /* ANDROID_X86_LINKER */
+
+#ifdef ANDROID_ARM_LINKER
+ case R_ARM_COPY:
+ COUNT_RELOC(RELOC_COPY);
+ MARK(rel->r_offset);
+ TRACE_TYPE(RELO, "%5d RELO %08x <- %d @ %08x %s\n", apkenv_pid,
+ reloc, s->st_size, sym_addr, sym_name);
+ memcpy((void*)reloc, (void*)sym_addr, s->st_size);
+ break;
+ case R_ARM_NONE:
+ break;
+#endif /* ANDROID_ARM_LINKER */
+
+ default:
+ DL_ERR("%5d unknown reloc type %d @ %p (%d)",
+ apkenv_pid, type, rel, (int) (rel - start));
+ return -1;
+ }
+ rel++;
+ }
+ return 0;
+}
+
+/* Please read the "Initialization and Termination functions" functions.
+ * of the linker design note in bionic/linker/README.TXT to understand
+ * what the following code is doing.
+ *
+ * The important things to remember are:
+ *
+ * DT_PREINIT_ARRAY must be called first for executables, and should
+ * not appear in shared libraries.
+ *
+ * DT_INIT should be called before DT_INIT_ARRAY if both are present
+ *
+ * DT_FINI should be called after DT_FINI_ARRAY if both are present
+ *
+ * DT_FINI_ARRAY must be parsed in reverse order.
+ */
+
+static void apkenv_call_array(unsigned *ctor, int count, int reverse)
+{
+ int n, inc = 1;
+
+ if (reverse) {
+ ctor += (count-1);
+ inc = -1;
+ }
+
+ for(n = count; n > 0; n--) {
+ TRACE("[ %5d Looking at %s *0x%08x == 0x%08x ]\n", apkenv_pid,
+ reverse ? "dtor" : "ctor",
+ (unsigned)ctor, (unsigned)*ctor);
+ void (*func)() = (void (*)()) *ctor;
+ ctor += inc;
+ if(((int) func == 0) || ((int) func == -1)) continue;
+ TRACE("[ %5d Calling func @ 0x%08x ]\n", apkenv_pid, (unsigned)func);
+ func();
+ }
+}
+
+void apkenv_call_constructors_recursive(soinfo *si)
+{
+ if (si->constructors_called)
+ return;
+ if (strcmp(si->name, "libc.so") == 0)
+return;
+ // Set this before actually calling the constructors, otherwise it doesn't
+ // protect against recursive constructor calls. One simple example of
+ // constructor recursion is the libc debug malloc, which is implemented in
+ // libc_malloc_debug_leak.so:
+ // 1. The program depends on libc, so libc's constructor is called here.
+ // 2. The libc constructor calls dlopen() to load libc_malloc_debug_leak.so.
+ // 3. dlopen() calls apkenv_call_constructors_recursive() with the newly created
+ // soinfo for libc_malloc_debug_leak.so.
+ // 4. The debug so depends on libc, so apkenv_call_constructors_recursive() is
+ // called again with the libc soinfo. If it doesn't trigger the early-
+ // out above, the libc constructor will be called again (recursively!).
+ si->constructors_called = 1;
+
+ if (si->flags & FLAG_EXE) {
+ TRACE("[ %5d Calling preinit_array @ 0x%08x [%d] for '%s' ]\n",
+ apkenv_pid, (unsigned)si->preinit_array, si->preinit_array_count,
+ si->name);
+ apkenv_call_array(si->preinit_array, si->preinit_array_count, 0);
+ TRACE("[ %5d Done calling preinit_array for '%s' ]\n", apkenv_pid, si->name);
+ } else {
+ if (si->preinit_array) {
+ DL_ERR("%5d Shared library '%s' has a preinit_array table @ 0x%08x."
+ " This is INVALID.", apkenv_pid, si->name,
+ (unsigned)si->preinit_array);
+ }
+ }
+
+ if (si->dynamic) {
+ unsigned *d;
+ for(d = si->dynamic; *d; d += 2) {
+ if(d[0] == DT_NEEDED){
+ soinfo* lsi = (soinfo *)d[1];
+ if (!apkenv_validate_soinfo(lsi)) {
+ DL_ERR("%5d bad DT_NEEDED pointer in %s",
+ apkenv_pid, si->name);
+ } else {
+ apkenv_call_constructors_recursive(lsi);
+ }
+ }
+ }
+ }
+
+ if (si->init_func) {
+ TRACE("[ %5d Calling init_func @ 0x%08x for '%s' ]\n", apkenv_pid,
+ (unsigned)si->init_func, si->name);
+ si->init_func();
+ TRACE("[ %5d Done calling init_func for '%s' ]\n", apkenv_pid, si->name);
+ }
+
+ if (si->init_array) {
+ TRACE("[ %5d Calling init_array @ 0x%08x [%d] for '%s' ]\n", apkenv_pid,
+ (unsigned)si->init_array, si->init_array_count, si->name);
+ apkenv_call_array(si->init_array, si->init_array_count, 0);
+ TRACE("[ %5d Done calling init_array for '%s' ]\n", apkenv_pid, si->name);
+ }
+
+}
+
+static void apkenv_call_destructors(soinfo *si)
+{
+ if (si->fini_array) {
+ TRACE("[ %5d Calling fini_array @ 0x%08x [%d] for '%s' ]\n", apkenv_pid,
+ (unsigned)si->fini_array, si->fini_array_count, si->name);
+ apkenv_call_array(si->fini_array, si->fini_array_count, 1);
+ TRACE("[ %5d Done calling fini_array for '%s' ]\n", apkenv_pid, si->name);
+ }
+
+ if (si->fini_func) {
+ TRACE("[ %5d Calling fini_func @ 0x%08x for '%s' ]\n", apkenv_pid,
+ (unsigned)si->fini_func, si->name);
+ si->fini_func();
+ TRACE("[ %5d Done calling fini_func for '%s' ]\n", apkenv_pid, si->name);
+ }
+}
+
+/* Force any of the closed stdin, stdout and stderr to be associated with
+ /dev/null. */
+static int apkenv_nullify_closed_stdio (void)
+{
+ int dev_null, i, status;
+ int return_value = 0;
+
+ dev_null = open("/dev/null", O_RDWR);
+ if (dev_null < 0) {
+ DL_ERR("Cannot open /dev/null.");
+ return -1;
+ }
+ TRACE("[ %5d Opened /dev/null file-descriptor=%d]\n", apkenv_pid, dev_null);
+
+ /* If any of the stdio file descriptors is valid and not associated
+ with /dev/null, dup /dev/null to it. */
+ for (i = 0; i < 3; i++) {
+ /* If it is /dev/null already, we are done. */
+ if (i == dev_null)
+ continue;
+
+ TRACE("[ %5d Nullifying stdio file descriptor %d]\n", apkenv_pid, i);
+ /* The man page of fcntl does not say that fcntl(..,F_GETFL)
+ can be interrupted but we do this just to be safe. */
+ do {
+ status = fcntl(i, F_GETFL);
+ } while (status < 0 && errno == EINTR);
+
+ /* If file is openned, we are good. */
+ if (status >= 0)
+ continue;
+
+ /* The only error we allow is that the file descriptor does not
+ exist, in which case we dup /dev/null to it. */
+ if (errno != EBADF) {
+ DL_ERR("nullify_stdio: unhandled error %s", strerror(errno));
+ return_value = -1;
+ continue;
+ }
+
+ /* Try dupping /dev/null to this stdio file descriptor and
+ repeat if there is a signal. Note that any errors in closing
+ the stdio descriptor are lost. */
+ do {
+ status = dup2(dev_null, i);
+ } while (status < 0 && errno == EINTR);
+
+ if (status < 0) {
+ DL_ERR("nullify_stdio: dup2 error %s", strerror(errno));
+ return_value = -1;
+ continue;
+ }
+ }
+
+ /* If /dev/null is not one of the stdio file descriptors, close it. */
+ if (dev_null > 2) {
+ TRACE("[ %5d Closing /dev/null file-descriptor=%d]\n", apkenv_pid, dev_null);
+ do {
+ status = close(dev_null);
+ } while (status < 0 && errno == EINTR);
+
+ if (status < 0) {
+ DL_ERR("nullify_stdio: close error %s", strerror(errno));
+ return_value = -1;
+ }
+ }
+
+ return return_value;
+}
+
+static void apkenv_wrap_function(void *sym_addr, char *sym_name, int is_thumb, soinfo *si)
+{
+ void *hook = NULL;
+#ifdef APKENV_LATEHOOKS
+ if((hook = apkenv_get_hooked_symbol(sym_name, 0)) != NULL)
+ {
+ // if we have a hook redirect the call to that by overwriting
+ // the first 2 instruction of the function
+ // this should work in any case unless the hook is wrong
+ // or the function shorter than 64-bit (2 32-Bit instructions)
+ if(!is_thumb)
+ {
+ DEBUG("HOOKING INTERNAL (ARM) FUNCTION %s@%x (in %s) TO: %x\n",sym_name,sym_addr,si->name,hook);
+ ((int32_t*)sym_addr)[0] = 0xe51ff004; // ldr pc, [pc, -#4] (load the hooks address into pc)
+ ((int32_t*)sym_addr)[1] = (uint32_t)create_wrapper(sym_name, hook, WRAPPER_LATEHOOK);
+
+ __clear_cache((int32_t*)sym_addr, (int32_t*)sym_addr + 2);
+ }
+ else
+ {
+ DEBUG("HOOKING INTERNAL (THUMB) FUNCTION %s@%x (in %s) TO: %x\n",sym_name,sym_addr,si->name,hook);
+ sym_addr = (void*)((char*)sym_addr - 1); // get actual sym_addr
+
+ ((int16_t*)sym_addr)[0] = 0xB401; /* push {r0} */
+ ((int16_t*)sym_addr)[1] = 0xF8DF; /* ldr r0, [pc, #8] */
+ ((int16_t*)sym_addr)[2] = 0x0008; /* continuation of last instruction */
+ ((int16_t*)sym_addr)[3] = 0x4684; /* mov ip, r0 */
+ ((int16_t*)sym_addr)[4] = 0xBC01; /* pop {r0} */
+ ((int16_t*)sym_addr)[5] = 0x4760; /* bx ip */
+
+ void *wrp = create_wrapper(sym_name, hook, WRAPPER_LATEHOOK);
+
+ // store the hooks address
+ ((int16_t*)sym_addr)[6] = (uint32_t)wrp & 0x0000FFFF;
+ ((int16_t*)sym_addr)[7] = (uint32_t)wrp >> 16;
+
+ __clear_cache((int16_t*)sym_addr, (int16_t*)sym_addr + 8);
+ }
+ }
+ else
+#endif /* APKENV_LATEHOOKS */
+ // TODO: this will fail if the first 2 instructions do something pc related
+ // (this DOES NOT happen very often)
+ if(sym_addr && !is_thumb)
+ {
+ DEBUG("CREATING ARM WRAPPER FOR: %s@%x (in %s)\n", sym_name, (unsigned)sym_addr, si->name);
+
+ create_wrapper(sym_name, sym_addr, WRAPPER_ARM_INJECTION);
+ }
+ // TODO: this will fail if the first 2-5 instructions do something pc related
+ // (this DOES happen very often)
+ else if(sym_addr && is_thumb)
+ {
+ DEBUG("CREATING THUMB WRAPPER FOR: %s@%x (in %s)\n",sym_name, (unsigned)sym_addr,si->name);
+
+ create_wrapper(sym_name, sym_addr, WRAPPER_THUMB_INJECTION);
+ }
+}
+
+
+/* Franz-Josef Haider apkenv_create_latehook_wrappers */
+static void apkenv_create_latehook_wrappers(soinfo *si)
+{
+ Elf32_Sym *s;
+
+ unsigned int n = 0;
+ for(n=0;n<si->nchain;n++)
+ {
+ s = si->symtab + n;
+ switch(ELF32_ST_BIND(s->st_info))
+ {
+ case STB_GLOBAL:
+ case STB_WEAK:
+ if(s->st_shndx == 0) continue;
+
+ if(ELF32_ST_TYPE(s->st_info) == STT_FUNC) // only wrap functions
+ {
+ char *sym_name = (char*)(si->strtab + s->st_name);
+ void *sym_addr = (void*)(intptr_t)(si->base + s->st_value);
+
+ int is_thumb = ((int32_t)(intptr_t)sym_addr) & 0x00000001;
+
+ apkenv_wrap_function(sym_addr,sym_name,is_thumb,si);
+ }
+ }
+ }
+}
+
+
+static int apkenv_link_image(soinfo *si, unsigned wr_offset)
+{
+ unsigned *d;
+ Elf32_Phdr *phdr = si->phdr;
+ int phnum = si->phnum;
+
+ INFO("[ %5d linking %s ]\n", apkenv_pid, si->name);
+ DEBUG("%5d si->base = 0x%08x si->flags = 0x%08x\n", apkenv_pid,
+ si->base, si->flags);
+
+ if (si->flags & (FLAG_EXE | FLAG_LINKER)) {
+ /* Locate the needed program segments (DYNAMIC/ARM_EXIDX) for
+ * linkage info if this is the executable or the linker itself.
+ * If this was a dynamic lib, that would have been done at load time.
+ *
+ * TODO: It's unfortunate that small pieces of this are
+ * repeated from the apkenv_load_library routine. Refactor this just
+ * slightly to reuse these bits.
+ */
+ si->size = 0;
+ for(; phnum > 0; --phnum, ++phdr) {
+#ifdef ANDROID_ARM_LINKER
+ if(phdr->p_type == PT_ARM_EXIDX) {
+ /* exidx entries (used for stack unwinding) are 8 bytes each.
+ */
+ si->ARM_exidx = (unsigned *)phdr->p_vaddr;
+ si->ARM_exidx_count = phdr->p_memsz / 8;
+ }
+#endif
+ if (phdr->p_type == PT_LOAD) {
+ /* For the executable, we use the si->size field only in
+ dl_unwind_find_exidx(), so the meaning of si->size
+ is not the size of the executable; it is the distance
+ between the load location of the executable and the last
+ address of the loadable part of the executable.
+ We use the range [si->base, si->base + si->size) to
+ determine whether a PC value falls within the executable
+ section. Of course, if a value is between si->base and
+ (si->base + phdr->p_vaddr), it's not in the executable
+ section, but a) we shouldn't be asking for such a value
+ anyway, and b) if we have to provide an EXIDX for such a
+ value, then the executable's EXIDX is probably the better
+ choice.
+ */
+ DEBUG_DUMP_PHDR(phdr, "PT_LOAD", apkenv_pid);
+ if (phdr->p_vaddr + phdr->p_memsz > si->size)
+ si->size = phdr->p_vaddr + phdr->p_memsz;
+ /* try to remember what range of addresses should be write
+ * protected */
+ if (!(phdr->p_flags & PF_W)) {
+ unsigned _end;
+
+ if (si->base + phdr->p_vaddr < si->wrprotect_start)
+ si->wrprotect_start = si->base + phdr->p_vaddr;
+ _end = (((si->base + phdr->p_vaddr + phdr->p_memsz + PAGE_SIZE - 1) &
+ (~PAGE_MASK)));
+ if (_end > si->wrprotect_end)
+ si->wrprotect_end = _end;
+ /* Make the section writable just in case we'll have to
+ * write to it during relocation (i.e. text segment).
+ * However, we will remember what range of addresses
+ * should be write protected.
+ */
+ mprotect((void *) (si->base + phdr->p_vaddr),
+ phdr->p_memsz,
+ PFLAGS_TO_PROT(phdr->p_flags) | PROT_WRITE);
+ }
+ } else if (phdr->p_type == PT_DYNAMIC) {
+ if (si->dynamic != (unsigned *)-1) {
+ DL_ERR("%5d multiple PT_DYNAMIC segments found in '%s'. "
+ "Segment at 0x%08x, previously one found at 0x%08x",
+ apkenv_pid, si->name, si->base + phdr->p_vaddr,
+ (unsigned)si->dynamic);
+ goto fail;
+ }
+ DEBUG_DUMP_PHDR(phdr, "PT_DYNAMIC", apkenv_pid);
+ si->dynamic = (unsigned *) (si->base + phdr->p_vaddr);
+ } else if (phdr->p_type == PT_GNU_RELRO) {
+ if ((phdr->p_vaddr >= si->size)
+ || ((phdr->p_vaddr + phdr->p_memsz) > si->size)
+ || ((si->base + phdr->p_vaddr + phdr->p_memsz) < si->base)) {
+ DL_ERR("%d invalid GNU_RELRO in '%s' "
+ "p_vaddr=0x%08x p_memsz=0x%08x", apkenv_pid, si->name,
+ phdr->p_vaddr, phdr->p_memsz);
+ goto fail;
+ }
+ si->gnu_relro_start = (Elf32_Addr) (si->base + phdr->p_vaddr);
+ si->gnu_relro_len = (unsigned) phdr->p_memsz;
+ }
+ }
+ }
+
+ if (si->dynamic == (unsigned *)-1) {
+ DL_ERR("%5d missing PT_DYNAMIC?!", apkenv_pid);
+ goto fail;
+ }
+
+ DEBUG("%5d dynamic = %p\n", apkenv_pid, si->dynamic);
+
+ /* extract useful information from dynamic section */
+ for(d = si->dynamic; *d; d++){
+ DEBUG("%5d d = %p, d[0] = 0x%08x d[1] = 0x%08x\n", apkenv_pid, d, d[0], d[1]);
+ switch(*d++){
+ case DT_HASH:
+ si->nbucket = ((unsigned *) (si->base + *d))[0];
+ si->nchain = ((unsigned *) (si->base + *d))[1];
+ si->bucket = (unsigned *) (si->base + *d + 8);
+ si->chain = (unsigned *) (si->base + *d + 8 + si->nbucket * 4);
+ break;
+ case DT_STRTAB:
+ si->strtab = (const char *) (si->base + *d);
+ break;
+ case DT_SYMTAB:
+ si->symtab = (Elf32_Sym *) (si->base + *d);
+ break;
+ case DT_PLTREL:
+ if(*d != DT_REL) {
+ DL_ERR("DT_RELA not supported");
+ goto fail;
+ }
+ break;
+ case DT_JMPREL:
+ si->plt_rel = (Elf32_Rel*) (si->base + *d);
+ break;
+ case DT_PLTRELSZ:
+ si->plt_rel_count = *d / 8;
+ break;
+ case DT_REL:
+ si->rel = (Elf32_Rel*) (si->base + *d);
+ break;
+ case DT_RELSZ:
+ si->rel_count = *d / 8;
+ break;
+ case DT_PLTGOT:
+ /* Save this in case we decide to do lazy binding. We don't yet. */
+ si->plt_got = (unsigned *)(si->base + *d);
+ break;
+ case DT_DEBUG:
+ // Set the DT_DEBUG entry to the addres of _r_debug for GDB
+ *d = (int) &_r_debug;
+ break;
+ case DT_RELA:
+ DL_ERR("%5d DT_RELA not supported", apkenv_pid);
+ goto fail;
+ case DT_INIT:
+ si->init_func = (void (*)(void))(si->base + *d);
+ DEBUG("%5d %s constructors (init func) found at %p\n",
+ apkenv_pid, si->name, si->init_func);
+ break;
+ case DT_FINI:
+ si->fini_func = (void (*)(void))(si->base + *d);
+ DEBUG("%5d %s destructors (fini func) found at %p\n",
+ apkenv_pid, si->name, si->fini_func);
+ break;
+ case DT_INIT_ARRAY:
+ si->init_array = (unsigned *)(si->base + *d);
+ DEBUG("%5d %s constructors (init_array) found at %p\n",
+ apkenv_pid, si->name, si->init_array);
+ break;
+ case DT_INIT_ARRAYSZ:
+ si->init_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
+ break;
+ case DT_FINI_ARRAY:
+ si->fini_array = (unsigned *)(si->base + *d);
+ DEBUG("%5d %s destructors (fini_array) found at %p\n",
+ apkenv_pid, si->name, si->fini_array);
+ break;
+ case DT_FINI_ARRAYSZ:
+ si->fini_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
+ break;
+ case DT_PREINIT_ARRAY:
+ si->preinit_array = (unsigned *)(si->base + *d);
+ DEBUG("%5d %s constructors (preinit_array) found at %p\n",
+ apkenv_pid, si->name, si->preinit_array);
+ break;
+ case DT_PREINIT_ARRAYSZ:
+ si->preinit_array_count = ((unsigned)*d) / sizeof(Elf32_Addr);
+ break;
+ case DT_TEXTREL:
+ /* TODO: make use of this. */
+ /* this means that we might have to write into where the text
+ * segment was loaded during relocation... Do something with
+ * it.
+ */
+ DEBUG("%5d Text segment should be writable during relocation.\n",
+ apkenv_pid);
+ break;
+ }
+ }
+
+ DEBUG("%5d si->base = 0x%08x, si->strtab = %p, si->symtab = %p\n",
+ apkenv_pid, si->base, si->strtab, si->symtab);
+
+ if((si->strtab == 0) || (si->symtab == 0)) {
+ DL_ERR("%5d missing essential tables", apkenv_pid);
+ goto fail;
+ }
+
+ /* if this is the main executable, then load all of the apkenv_preloads now */
+ if(si->flags & FLAG_EXE) {
+ int i;
+ memset(apkenv_preloads, 0, sizeof(apkenv_preloads));
+ for(i = 0; apkenv_ldpreload_names[i] != NULL; i++) {
+ soinfo *lsi = apkenv_find_library(apkenv_ldpreload_names[i]);
+ if(lsi == 0) {
+ apkenv_strlcpy(apkenv_tmp_err_buf, apkenv_linker_get_error(), sizeof(apkenv_tmp_err_buf));
+ DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
+ apkenv_pid, apkenv_ldpreload_names[i], si->name, apkenv_tmp_err_buf);
+ goto fail;
+ }
+ lsi->refcount++;
+ apkenv_preloads[i] = lsi;
+ }
+ }
+
+ for(d = si->dynamic; *d; d += 2) {
+ if(d[0] == DT_NEEDED){
+ DEBUG("%5d %s needs %s\n", apkenv_pid, si->name, si->strtab + d[1]);
+ soinfo *lsi = NULL;
+#if 1
+ // if (get_builtin_lib_handle(si->strtab + d[1]) == NULL)
+ lsi = apkenv_find_library(si->strtab + d[1]);
+ if (!lsi && dlopen(si->strtab + d[1], RTLD_NOW | RTLD_GLOBAL)) {
+ DEBUG("Loaded %s with glibc dlopen\n", si->strtab + d[1]);
+ }
+#endif
+ if(lsi == 0) {
+ /**
+ * XXX Dirty Hack Alarm --thp XXX
+ *
+ * For libraries that are not found, simply use libdl's soinfo,
+ * in order to not break downstream users of the library while
+ * still allowing the application to start up.
+ **/
+ lsi = &apkenv_libdl_info;
+ //apkenv_strlcpy(apkenv_tmp_err_buf, apkenv_linker_get_error(), sizeof(apkenv_tmp_err_buf));
+ //DL_ERR("%5d could not load needed library '%s' for '%s' (%s)",
+ // apkenv_pid, si->strtab + d[1], si->name, apkenv_tmp_err_buf);
+ //continue;
+ //goto fail;
+ }
+ /* Save the soinfo of the loaded DT_NEEDED library in the payload
+ of the DT_NEEDED entry itself, so that we can retrieve the
+ soinfo directly later from the dynamic segment. This is a hack,
+ but it allows us to map from DT_NEEDED to soinfo efficiently
+ later on when we resolve relocations, trying to look up a symbol
+ with dlsym().
+ */
+ d[1] = (intptr_t)lsi;
+ lsi->refcount++;
+ }
+ }
+
+ if(si->plt_rel) {
+ DEBUG("[ %5d relocating %s plt ]\n", apkenv_pid, si->name );
+ if(apkenv_reloc_library(si, si->plt_rel, si->plt_rel_count))
+ goto fail;
+ }
+ if(si->rel) {
+ DEBUG("[ %5d relocating %s ]\n", apkenv_pid, si->name );
+ if(apkenv_reloc_library(si, si->rel, si->rel_count))
+ goto fail;
+ }
+
+ apkenv_create_latehook_wrappers(si);
+
+ si->flags |= FLAG_LINKED;
+ DEBUG("[ %5d finished linking %s ]\n", apkenv_pid, si->name);
+
+#if 0
+ /* This is the way that the old dynamic linker did protection of
+ * non-writable areas. It would scan section headers and find where
+ * .text ended (rather where .data/.bss began) and assume that this is
+ * the upper range of the non-writable area. This is too coarse,
+ * and is kept here for reference until we fully move away from single
+ * segment elf objects. See the code in get_wr_offset (also #if'd 0)
+ * that made this possible.
+ */
+ if(wr_offset < 0xffffffff){
+ mprotect((void*) si->base, wr_offset, PROT_READ | PROT_EXEC);
+ }
+#else
+ /* TODO: Verify that this does the right thing in all cases, as it
+ * presently probably does not. It is possible that an ELF image will
+ * come with multiple read-only segments. What we ought to do is scan
+ * the program headers again and mprotect all the read-only segments.
+ * To prevent re-scanning the program header, we would have to build a
+ * list of loadable segments in si, and then scan that instead. */
+ if (si->wrprotect_start != 0xffffffff && si->wrprotect_end != 0) {
+ mprotect((void *)(intptr_t)si->wrprotect_start,
+ si->wrprotect_end - si->wrprotect_start,
+ PROT_READ | PROT_EXEC);
+ }
+#endif
+
+ if (si->gnu_relro_start != 0 && si->gnu_relro_len != 0) {
+ Elf32_Addr start = (si->gnu_relro_start & ~PAGE_MASK);
+ unsigned len = (si->gnu_relro_start - start) + si->gnu_relro_len;
+ if (mprotect((void *)(intptr_t)start, len, PROT_READ) < 0) {
+ DL_ERR("%5d GNU_RELRO mprotect of library '%s' failed: %d (%s)\n",
+ apkenv_pid, si->name, errno, strerror(errno));
+ goto fail;
+ }
+ }
+
+ /* If this is a SET?ID program, dup /dev/null to opened stdin,
+ stdout and stderr to close a security hole described in:
+
+ ftp://ftp.freebsd.org/pub/FreeBSD/CERT/advisories/FreeBSD-SA-02:23.stdio.asc
+
+ */
+ if (apkenv_program_is_setuid)
+ apkenv_nullify_closed_stdio ();
+ apkenv_notify_gdb_of_load(si);
+ return 0;
+
+fail:
+ ERROR("failed to link %s\n", si->name);
+ si->flags |= FLAG_ERROR;
+ return -1;
+}
+
+void apkenv_parse_library_path(const char *path, char *delim)
+{
+ size_t len;
+ char *apkenv_ldpaths_bufp = apkenv_ldpaths_buf;
+ int i = 0;
+
+ len = apkenv_strlcpy(apkenv_ldpaths_buf, path, sizeof(apkenv_ldpaths_buf));
+
+ while (i < LDPATH_MAX && (apkenv_ldpaths[i] = strsep(&apkenv_ldpaths_bufp, delim))) {
+ if (*apkenv_ldpaths[i] != '\0')
+ ++i;
+ }
+
+ /* Forget the last path if we had to truncate; this occurs if the 2nd to
+ * last char isn't '\0' (i.e. not originally a delim). */
+ if (i > 0 && len >= sizeof(apkenv_ldpaths_buf) &&
+ apkenv_ldpaths_buf[sizeof(apkenv_ldpaths_buf) - 2] != '\0') {
+ apkenv_ldpaths[i - 1] = NULL;
+ } else {
+ apkenv_ldpaths[i] = NULL;
+ }
+}
+
+static void apkenv_parse_apkenv_preloads(const char *path, char *delim)
+{
+ size_t len;
+ char *apkenv_ldpreloads_bufp = apkenv_ldpreloads_buf;
+ int i = 0;
+
+ len = apkenv_strlcpy(apkenv_ldpreloads_buf, path, sizeof(apkenv_ldpreloads_buf));
+
+ while (i < LDPRELOAD_MAX && (apkenv_ldpreload_names[i] = strsep(&apkenv_ldpreloads_bufp, delim))) {
+ if (*apkenv_ldpreload_names[i] != '\0') {
+ ++i;
+ }
+ }
+
+ /* Forget the last path if we had to truncate; this occurs if the 2nd to
+ * last char isn't '\0' (i.e. not originally a delim). */
+ if (i > 0 && len >= sizeof(apkenv_ldpreloads_buf) &&
+ apkenv_ldpreloads_buf[sizeof(apkenv_ldpreloads_buf) - 2] != '\0') {
+ apkenv_ldpreload_names[i - 1] = NULL;
+ } else {
+ apkenv_ldpreload_names[i] = NULL;
+ }
+}
+
+/*
+ * This code is called after the linker has linked itself and
+ * fixed it's own GOT. It is safe to make references to externs
+ * and other non-local data at this point.
+ */
+static unsigned apkenv___linker_init_post_relocation(unsigned **elfdata)
+{
+ static soinfo linker_soinfo;
+
+ int argc = (int)(intptr_t)*elfdata;
+ char **argv = (char**) (elfdata + 1);
+ unsigned *vecs = (unsigned*) (argv + argc + 1);
+ unsigned *v;
+ soinfo *si;
+ struct link_map * map;
+ const char *ldpath_env = NULL;
+ const char *ldpreload_env = NULL;
+
+ /* NOTE: we store the elfdata pointer on a special location
+ * of the temporary TLS area in order to pass it to
+ * the C Library's runtime initializer.
+ *
+ * The initializer must clear the slot and reset the TLS
+ * to point to a different location to ensure that no other
+ * shared library constructor can access it.
+ */
+#if 0
+ __libc_init_tls(elfdata);
+#endif
+
+ apkenv_pid = getpid();
+
+#if TIMING
+ struct timeval t0, t1;
+ gettimeofday(&t0, 0);
+#endif
+
+ /* Initialize environment functions, and get to the ELF aux vectors table */
+ vecs = apkenv_linker_env_init(vecs);
+
+ /* Check auxv for AT_SECURE first to see if program is setuid, setgid,
+ has file caps, or caused a SELinux/AppArmor domain transition. */
+ for (v = vecs; v[0]; v += 2) {
+ if (v[0] == AT_SECURE) {
+ /* kernel told us whether to enable secure mode */
+ apkenv_program_is_setuid = v[1];
+ goto sanitize;
+ }
+ }
+
+ /* Kernel did not provide AT_SECURE - fall back on legacy test. */
+ apkenv_program_is_setuid = (getuid() != geteuid()) || (getgid() != getegid());
+
+sanitize:
+ /* Sanitize environment if we're loading a setuid program */
+ if (apkenv_program_is_setuid)
+ apkenv_linker_env_secure();
+
+#if 0
+ apkenv_debugger_init();
+#endif
+
+ /* Get a few environment variables */
+ {
+#if LINKER_DEBUG
+ const char* env;
+ env = apkenv_linker_env_get("DEBUG"); /* XXX: TODO: Change to LD_DEBUG */
+ if (env)
+ apkenv_debug_verbosity = atoi(env);
+#endif
+
+ /* Normally, these are cleaned by linker_env_secure, but the test
+ * against apkenv_program_is_setuid doesn't cost us anything */
+ if (!apkenv_program_is_setuid) {
+ ldpath_env = getenv("LD_LIBRARY_PATH");
+ ldpreload_env = getenv("LD_PRELOAD");
+ }
+ }
+
+ INFO("[ android linker & debugger ]\n");
+ DEBUG("%5d elfdata @ 0x%08x\n", apkenv_pid, (unsigned)elfdata);
+
+ si = apkenv_alloc_info(argv[0]);
+ if(si == 0) {
+ exit(-1);
+ }
+
+ /* bootstrap the link map, the main exe always needs to be first */
+ si->flags |= FLAG_EXE;
+ map = &(si->linkmap);
+
+ map->l_addr = 0;
+ map->l_name = argv[0];
+ map->l_prev = NULL;
+ map->l_next = NULL;
+
+ _r_debug.r_map = map;
+ apkenv_r_debug_tail = map;
+
+ /* gdb expects the linker to be in the debug shared object list,
+ * and we need to make sure that the reported load address is zero.
+ * Without this, gdb gets the wrong idea of where rtld_db_dlactivity()
+ * is. Don't use apkenv_alloc_info(), because the linker shouldn't
+ * be on the soinfo list.
+ */
+ apkenv_strlcpy((char*) linker_soinfo.name, "/system/bin/linker", sizeof linker_soinfo.name);
+ linker_soinfo.flags = 0;
+ linker_soinfo.base = 0; // This is the important part; must be zero.
+ apkenv_insert_soinfo_into_debug_map(&linker_soinfo);
+
+ /* extract information passed from the kernel */
+ while(vecs[0] != 0){
+ switch(vecs[0]){
+ case AT_PHDR:
+ si->phdr = (Elf32_Phdr*)(intptr_t)vecs[1];
+ break;
+ case AT_PHNUM:
+ si->phnum = (int) vecs[1];
+ break;
+ case AT_ENTRY:
+ si->entry = vecs[1];
+ break;
+ }
+ vecs += 2;
+ }
+
+ /* Compute the value of si->base. We can't rely on the fact that
+ * the first entry is the PHDR because this will not be true
+ * for certain executables (e.g. some in the NDK unit test suite)
+ */
+ int nn;
+ si->base = 0;
+ for ( nn = 0; nn < si->phnum; nn++ ) {
+ if (si->phdr[nn].p_type == PT_PHDR) {
+ si->base = (Elf32_Addr)(intptr_t)si->phdr - si->phdr[nn].p_vaddr;
+ break;
+ }
+ }
+ si->dynamic = (unsigned *)-1;
+ si->wrprotect_start = 0xffffffff;
+ si->wrprotect_end = 0;
+ si->refcount = 1;
+ si->gnu_relro_start = 0;
+ si->gnu_relro_len = 0;
+
+ /* Use LD_LIBRARY_PATH if we aren't setuid/setgid */
+ if (ldpath_env)
+ apkenv_parse_library_path(ldpath_env, ":");
+
+ if (ldpreload_env) {
+ apkenv_parse_apkenv_preloads(ldpreload_env, " :");
+ }
+ if(apkenv_link_image(si, 0)) {
+ char errmsg[] = "CANNOT LINK EXECUTABLE\n";
+ write(2, apkenv___linker_dl_err_buf, strlen(apkenv___linker_dl_err_buf));
+ write(2, errmsg, sizeof(errmsg));
+ exit(-1);
+ }
+
+ apkenv_call_constructors_recursive(si);
+
+#if ALLOW_SYMBOLS_FROM_MAIN
+ /* Set apkenv_somain after we've loaded all the libraries in order to prevent
+ * linking of symbols back to the main image, which is not set up at that
+ * point yet.
+ */
+ apkenv_somain = si;
+#endif
+
+#if TIMING
+ gettimeofday(&t1,NULL);
+ PRINT("LINKER TIME: %s: %d microseconds\n", argv[0], (int) (
+ (((long long)t1.tv_sec * 1000000LL) + (long long)t1.tv_usec) -
+ (((long long)t0.tv_sec * 1000000LL) + (long long)t0.tv_usec)
+ ));
+#endif
+#if STATS
+ PRINT("RELO STATS: %s: %d abs, %d rel, %d copy, %d symbol\n", argv[0],
+ apkenv_linker_stats.reloc[RELOC_ABSOLUTE],
+ apkenv_linker_stats.reloc[RELOC_RELATIVE],
+ apkenv_linker_stats.reloc[RELOC_COPY],
+ apkenv_linker_stats.reloc[RELOC_SYMBOL]);
+#endif
+#if COUNT_PAGES
+ {
+ unsigned n;
+ unsigned i;
+ unsigned count = 0;
+ for(n = 0; n < 4096; n++){
+ if(apkenv_bitmask[n]){
+ unsigned x = apkenv_bitmask[n];
+ for(i = 0; i < 8; i++){
+ if(x & 1) count++;
+ x >>= 1;
+ }
+ }
+ }
+ PRINT("PAGES MODIFIED: %s: %d (%dKB)\n", argv[0], count, count * 4);
+ }
+#endif
+
+#if TIMING || STATS || COUNT_PAGES
+ fflush(stdout);
+#endif
+
+ TRACE("[ %5d Ready to execute '%s' @ 0x%08x ]\n", apkenv_pid, si->name,
+ si->entry);
+ return si->entry;
+}
+
+/*
+ * Find the value of AT_BASE passed to us by the kernel. This is the load
+ * location of the linker.
+ */
+static unsigned apkenv_find_linker_base(unsigned **elfdata) {
+ int argc = (int)(intptr_t)*elfdata;
+ char **argv = (char**) (elfdata + 1);
+ unsigned *vecs = (unsigned*) (argv + argc + 1);
+ while (vecs[0] != 0) {
+ vecs++;
+ }
+
+ /* The end of the environment block is marked by two NULL pointers */
+ vecs++;
+
+ while(vecs[0]) {
+ if (vecs[0] == AT_BASE) {
+ return vecs[1];
+ }
+ vecs += 2;
+ }
+
+ return 0; // should never happen
+}
+
+/*
+ * This is the entry point for the linker, called from begin.S. This
+ * method is responsible for fixing the linker's own relocations, and
+ * then calling apkenv___linker_init_post_relocation().
+ *
+ * Because this method is called before the linker has fixed it's own
+ * relocations, any attempt to reference an extern variable, extern
+ * function, or other GOT reference will generate a segfault.
+ */
+unsigned apkenv___linker_init(unsigned **elfdata) {
+ unsigned linker_addr = apkenv_find_linker_base(elfdata);
+ Elf32_Ehdr *elf_hdr = (Elf32_Ehdr *)(intptr_t)linker_addr;
+ Elf32_Phdr *phdr =
+ (Elf32_Phdr *)((unsigned char *)(intptr_t)linker_addr + elf_hdr->e_phoff);
+
+ soinfo linker_so;
+ memset(&linker_so, 0, sizeof(soinfo));
+
+ linker_so.base = linker_addr;
+ linker_so.dynamic = (unsigned *) -1;
+ linker_so.phdr = phdr;
+ linker_so.phnum = elf_hdr->e_phnum;
+ linker_so.flags |= FLAG_LINKER;
+ linker_so.wrprotect_start = 0xffffffff;
+ linker_so.wrprotect_end = 0;
+ linker_so.gnu_relro_start = 0;
+ linker_so.gnu_relro_len = 0;
+ if (apkenv_link_image(&linker_so, 0)) {
+ // It would be nice to print an error message, but if the linker
+ // can't link itself, there's no guarantee that we'll be able to
+ // call write() (because it involves a GOT reference).
+ //
+ // This situation should never occur unless the linker itself
+ // is corrupt.
+ exit(-1);
+ }
+
+ // We have successfully fixed our own relocations. It's safe to run
+ // the main part of the linker now.
+ return apkenv___linker_init_post_relocation(elfdata);
+}