// Copyright 2014 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "components/nacl/loader/nonsfi/elf_loader.h" #include #include #include #include #include #include "base/logging.h" #include "base/strings/string_number_conversions.h" #include "native_client/src/include/portability.h" #include "native_client/src/trusted/desc/nacl_desc_base.h" #include "native_client/src/trusted/desc/nacl_desc_effector_trusted_mem.h" #include "native_client/src/trusted/service_runtime/include/bits/mman.h" // Extracted from native_client/src/trusted/service_runtime/nacl_config.h. #if NACL_ARCH(NACL_BUILD_ARCH) == NACL_x86 # if NACL_BUILD_SUBARCH == 64 # define NACL_ELF_E_MACHINE EM_X86_64 # elif NACL_BUILD_SUBARCH == 32 # define NACL_ELF_E_MACHINE EM_386 # else # error Unknown platform. # endif #elif NACL_ARCH(NACL_BUILD_ARCH) == NACL_arm # define NACL_ELF_E_MACHINE EM_ARM #elif NACL_ARCH(NACL_BUILD_ARCH) == NACL_mips # define NACL_ELF_E_MACHINE EM_MIPS #else # error Unknown platform. #endif namespace nacl { namespace nonsfi { namespace { // Page size for non-SFI Mode. const ElfW(Addr) kNonSfiPageSize = 4096; const ElfW(Addr) kNonSfiPageMask = kNonSfiPageSize - 1; NaClErrorCode ValidateElfHeader(const ElfW(Ehdr)& ehdr) { if (std::memcmp(ehdr.e_ident, ELFMAG, SELFMAG)) { LOG(ERROR) << "Bad elf magic"; return LOAD_BAD_ELF_MAGIC; } #if NACL_BUILD_SUBARCH == 32 if (ehdr.e_ident[EI_CLASS] != ELFCLASS32) { LOG(ERROR) << "Bad elf class"; return LOAD_NOT_32_BIT; } #elif NACL_BUILD_SUBARCH == 64 if (ehdr.e_ident[EI_CLASS] != ELFCLASS64) { LOG(ERROR) << "Bad elf class"; return LOAD_NOT_64_BIT; } #else # error Unknown platform. #endif if (ehdr.e_type != ET_DYN) { LOG(ERROR) << "Not a relocatable ELF object (not ET_DYN)"; return LOAD_NOT_EXEC; } if (ehdr.e_machine != NACL_ELF_E_MACHINE) { LOG(ERROR) << "Bad machine: " << base::HexEncode(&ehdr.e_machine, sizeof(ehdr.e_machine)); return LOAD_BAD_MACHINE; } if (ehdr.e_version != EV_CURRENT) { LOG(ERROR) << "Bad elf version: " << base::HexEncode(&ehdr.e_version, sizeof(ehdr.e_version)); } return LOAD_OK; } // Returns the address of the page starting at address 'addr' for non-SFI mode. ElfW(Addr) GetPageStart(ElfW(Addr) addr) { return addr & ~kNonSfiPageMask; } // Returns the offset of address 'addr' in its memory page. In other words, // this equals to 'addr' - GetPageStart(addr). ElfW(Addr) GetPageOffset(ElfW(Addr) addr) { return addr & kNonSfiPageMask; } // Returns the address of the next page after address 'addr', unless 'addr' is // at the start of a page. This equals to: // addr == GetPageStart(addr) ? addr : GetPageStart(addr) + kNonSfiPageSize ElfW(Addr) GetPageEnd(ElfW(Addr) addr) { return GetPageStart(addr + kNonSfiPageSize - 1); } // Converts the pflags (in phdr) to mmap's prot flags. int PFlagsToProt(int pflags) { return ((pflags & PF_X) ? PROT_EXEC : 0) | ((pflags & PF_R) ? PROT_READ : 0) | ((pflags & PF_W) ? PROT_WRITE : 0); } // Converts the pflags (in phdr) to NaCl ABI's prot flags. int PFlagsToNaClProt(int pflags) { return ((pflags & PF_X) ? NACL_ABI_PROT_EXEC : 0) | ((pflags & PF_R) ? NACL_ABI_PROT_READ : 0) | ((pflags & PF_W) ? NACL_ABI_PROT_WRITE : 0); } // Returns the load size for the given phdrs, or 0 on error. ElfW(Addr) GetLoadSize(const ElfW(Phdr)* phdrs, int phnum) { ElfW(Addr) begin = ~static_cast(0); ElfW(Addr) end = 0; for (int i = 0; i < phnum; ++i) { const ElfW(Phdr)& phdr = phdrs[i]; if (phdr.p_type != PT_LOAD) { // Do nothing for non PT_LOAD header. continue; } begin = std::min(begin, phdr.p_vaddr); end = std::max(end, phdr.p_vaddr + phdr.p_memsz); } if (begin > end) { // The end address looks overflowing, or PT_LOAD is not found. return 0; } return GetPageEnd(end) - GetPageStart(begin); } // Reserves the memory for the given phdrs, and stores the memory bias to the // load_bias. NaClErrorCode ReserveMemory(const ElfW(Phdr)* phdrs, int phnum, ElfW(Addr)* load_bias) { ElfW(Addr) size = GetLoadSize(phdrs, phnum); if (size == 0) { LOG(ERROR) << "ReserveMemory failed to calculate size"; return LOAD_UNLOADABLE; } // Make sure that the given program headers represents PIE binary. for (int i = 0; i < phnum; ++i) { if (phdrs[i].p_type == PT_LOAD) { // Here, phdrs[i] is the first loadable segment. if (phdrs[i].p_vaddr != 0) { // The binary is not PIE (i.e. needs to be loaded onto fixed addressed // memory. We don't support such a case. LOG(ERROR) << "ReserveMemory: Non-PIE binary loading is not supported."; return LOAD_UNLOADABLE; } break; } } void* start = mmap(0, size, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); if (start == MAP_FAILED) { LOG(ERROR) << "ReserveMemory: failed to mmap."; return LOAD_NO_MEMORY; } *load_bias = reinterpret_cast(start); return LOAD_OK; } NaClErrorCode LoadSegments( const ElfW(Phdr)* phdrs, int phnum, ElfW(Addr) load_bias, struct NaClDesc* descriptor) { for (int i = 0; i < phnum; ++i) { const ElfW(Phdr)& phdr = phdrs[i]; if (phdr.p_type != PT_LOAD) { // Not a load target. continue; } // Addresses on the memory. ElfW(Addr) seg_start = phdr.p_vaddr + load_bias; ElfW(Addr) seg_end = seg_start + phdr.p_memsz; ElfW(Addr) seg_page_start = GetPageStart(seg_start); ElfW(Addr) seg_page_end = GetPageEnd(seg_end); ElfW(Addr) seg_file_end = seg_start + phdr.p_filesz; // Addresses on the file content. ElfW(Addr) file_start = phdr.p_offset; ElfW(Addr) file_end = file_start + phdr.p_filesz; ElfW(Addr) file_page_start = GetPageStart(file_start); uintptr_t seg_addr = (*NACL_VTBL(NaClDesc, descriptor)->Map)( descriptor, NaClDescEffectorTrustedMem(), reinterpret_cast(seg_page_start), file_end - file_page_start, PFlagsToNaClProt(phdr.p_flags), NACL_ABI_MAP_PRIVATE | NACL_ABI_MAP_FIXED, file_page_start); if (NaClPtrIsNegErrno(&seg_addr)) { LOG(ERROR) << "LoadSegments: [" << i << "] mmap failed, " << seg_addr; return LOAD_NO_MEMORY; } // Handle the BSS: fill Zero between the segment end and the page boundary // if necessary (i.e. if the segment doesn't end on a page boundary). ElfW(Addr) seg_file_end_offset = GetPageOffset(seg_file_end); if ((phdr.p_flags & PF_W) && seg_file_end_offset > 0) { memset(reinterpret_cast(seg_file_end), 0, kNonSfiPageSize - seg_file_end_offset); } // Hereafter, seg_file_end is now the first page address after the file // content. If seg_end is larger, we need to zero anything between them. // This is done by using a private anonymous mmap for all extra pages. seg_file_end = GetPageEnd(seg_file_end); if (seg_page_end > seg_file_end) { void* zeromap = mmap(reinterpret_cast(seg_file_end), seg_page_end - seg_file_end, PFlagsToProt(phdr.p_flags), MAP_FIXED | MAP_ANONYMOUS | MAP_PRIVATE, -1, 0); if (zeromap == MAP_FAILED) { LOG(ERROR) << "LoadSegments: [" << i << "] Failed to zeromap."; return LOAD_NO_MEMORY; } } } return LOAD_OK; } } // namespace struct ElfImage::Data { // Limit of elf program headers allowed. enum { MAX_PROGRAM_HEADERS = 128 }; ElfW(Ehdr) ehdr; ElfW(Phdr) phdrs[MAX_PROGRAM_HEADERS]; ElfW(Addr) load_bias; }; ElfImage::ElfImage() { } ElfImage::~ElfImage() { } uintptr_t ElfImage::entry_point() const { if (!data_) { LOG(DFATAL) << "entry_point must be called after Read()."; return 0; } return data_->ehdr.e_entry + data_->load_bias; } NaClErrorCode ElfImage::Read(struct NaClDesc* descriptor) { DCHECK(!data_); ::scoped_ptr data(new Data); // Read elf header. ssize_t read_ret = (*NACL_VTBL(NaClDesc, descriptor)->PRead)( descriptor, &data->ehdr, sizeof(data->ehdr), 0); if (NaClSSizeIsNegErrno(&read_ret) || static_cast(read_ret) != sizeof(data->ehdr)) { LOG(ERROR) << "Could not load elf headers."; return LOAD_READ_ERROR; } NaClErrorCode error_code = ValidateElfHeader(data->ehdr); if (error_code != LOAD_OK) return error_code; // Read program headers. if (data->ehdr.e_phnum > Data::MAX_PROGRAM_HEADERS) { LOG(ERROR) << "Too many program headers"; return LOAD_TOO_MANY_PROG_HDRS; } if (data->ehdr.e_phentsize != sizeof(data->phdrs[0])) { LOG(ERROR) << "Bad program headers size\n" << " ehdr_.e_phentsize = " << data->ehdr.e_phentsize << "\n" << " sizeof phdrs[0] = " << sizeof(data->phdrs[0]); return LOAD_BAD_PHENTSIZE; } size_t read_size = data->ehdr.e_phnum * data->ehdr.e_phentsize; read_ret = (*NACL_VTBL(NaClDesc, descriptor)->PRead)( descriptor, data->phdrs, read_size, data->ehdr.e_phoff); if (NaClSSizeIsNegErrno(&read_ret) || static_cast(read_ret) != read_size) { LOG(ERROR) << "Cannot load prog headers"; return LOAD_READ_ERROR; } data_.swap(data); return LOAD_OK; } NaClErrorCode ElfImage::Load(struct NaClDesc* descriptor) { if (!data_) { LOG(DFATAL) << "ElfImage::Load() must be called after Read()"; return LOAD_INTERNAL; } NaClErrorCode error = ReserveMemory(data_->phdrs, data_->ehdr.e_phnum, &data_->load_bias); if (error != LOAD_OK) { LOG(ERROR) << "ElfImage::Load: Failed to allocate memory"; return error; } error = LoadSegments( data_->phdrs, data_->ehdr.e_phnum, data_->load_bias, descriptor); if (error != LOAD_OK) { LOG(ERROR) << "ElfImage::Load: Failed to load segments"; return error; } return LOAD_OK; } } // namespace nonsfi } // namespace nacl