1353944Sdim//===-- fuchsia.cpp ---------------------------------------------*- C++ -*-===// 2353944Sdim// 3353944Sdim// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4353944Sdim// See https://llvm.org/LICENSE.txt for license information. 5353944Sdim// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6353944Sdim// 7353944Sdim//===----------------------------------------------------------------------===// 8353944Sdim 9353944Sdim#include "platform.h" 10353944Sdim 11353944Sdim#if SCUDO_FUCHSIA 12353944Sdim 13353944Sdim#include "common.h" 14353944Sdim#include "mutex.h" 15353944Sdim#include "string_utils.h" 16353944Sdim 17353944Sdim#include <lib/sync/mutex.h> // for sync_mutex_t 18353944Sdim#include <limits.h> // for PAGE_SIZE 19353944Sdim#include <stdlib.h> // for getenv() 20353944Sdim#include <zircon/compiler.h> 21353944Sdim#include <zircon/sanitizer.h> 22353944Sdim#include <zircon/syscalls.h> 23353944Sdim 24353944Sdimnamespace scudo { 25353944Sdim 26353944Sdimuptr getPageSize() { return PAGE_SIZE; } 27353944Sdim 28353944Sdimvoid NORETURN die() { __builtin_trap(); } 29353944Sdim 30353944Sdim// We zero-initialize the Extra parameter of map(), make sure this is consistent 31353944Sdim// with ZX_HANDLE_INVALID. 32357095Sdimstatic_assert(ZX_HANDLE_INVALID == 0, ""); 33353944Sdim 34353944Sdimstatic void *allocateVmar(uptr Size, MapPlatformData *Data, bool AllowNoMem) { 35353944Sdim // Only scenario so far. 36353944Sdim DCHECK(Data); 37353944Sdim DCHECK_EQ(Data->Vmar, ZX_HANDLE_INVALID); 38353944Sdim 39353944Sdim const zx_status_t Status = _zx_vmar_allocate( 40353944Sdim _zx_vmar_root_self(), 41353944Sdim ZX_VM_CAN_MAP_READ | ZX_VM_CAN_MAP_WRITE | ZX_VM_CAN_MAP_SPECIFIC, 0, 42353944Sdim Size, &Data->Vmar, &Data->VmarBase); 43353944Sdim if (UNLIKELY(Status != ZX_OK)) { 44353944Sdim if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) 45353944Sdim dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY); 46353944Sdim return nullptr; 47353944Sdim } 48353944Sdim return reinterpret_cast<void *>(Data->VmarBase); 49353944Sdim} 50353944Sdim 51353944Sdimvoid *map(void *Addr, uptr Size, const char *Name, uptr Flags, 52353944Sdim MapPlatformData *Data) { 53353944Sdim DCHECK_EQ(Size % PAGE_SIZE, 0); 54353944Sdim const bool AllowNoMem = !!(Flags & MAP_ALLOWNOMEM); 55353944Sdim 56353944Sdim // For MAP_NOACCESS, just allocate a Vmar and return. 57353944Sdim if (Flags & MAP_NOACCESS) 58353944Sdim return allocateVmar(Size, Data, AllowNoMem); 59353944Sdim 60353944Sdim const zx_handle_t Vmar = Data ? Data->Vmar : _zx_vmar_root_self(); 61353944Sdim CHECK_NE(Vmar, ZX_HANDLE_INVALID); 62353944Sdim 63353944Sdim zx_status_t Status; 64353944Sdim zx_handle_t Vmo; 65353944Sdim uint64_t VmoSize = 0; 66353944Sdim if (Data && Data->Vmo != ZX_HANDLE_INVALID) { 67353944Sdim // If a Vmo was specified, it's a resize operation. 68353944Sdim CHECK(Addr); 69353944Sdim DCHECK(Flags & MAP_RESIZABLE); 70353944Sdim Vmo = Data->Vmo; 71353944Sdim VmoSize = Data->VmoSize; 72353944Sdim Status = _zx_vmo_set_size(Vmo, VmoSize + Size); 73353944Sdim if (Status != ZX_OK) { 74353944Sdim if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) 75353944Sdim dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY); 76353944Sdim return nullptr; 77353944Sdim } 78353944Sdim } else { 79353944Sdim // Otherwise, create a Vmo and set its name. 80353944Sdim Status = _zx_vmo_create(Size, ZX_VMO_RESIZABLE, &Vmo); 81353944Sdim if (UNLIKELY(Status != ZX_OK)) { 82353944Sdim if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) 83353944Sdim dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY); 84353944Sdim return nullptr; 85353944Sdim } 86353944Sdim _zx_object_set_property(Vmo, ZX_PROP_NAME, Name, strlen(Name)); 87353944Sdim } 88353944Sdim 89353944Sdim uintptr_t P; 90353944Sdim zx_vm_option_t MapFlags = 91353944Sdim ZX_VM_PERM_READ | ZX_VM_PERM_WRITE | ZX_VM_ALLOW_FAULTS; 92353944Sdim const uint64_t Offset = 93353944Sdim Addr ? reinterpret_cast<uintptr_t>(Addr) - Data->VmarBase : 0; 94353944Sdim if (Offset) 95353944Sdim MapFlags |= ZX_VM_SPECIFIC; 96353944Sdim Status = _zx_vmar_map(Vmar, MapFlags, Offset, Vmo, VmoSize, Size, &P); 97353944Sdim // No need to track the Vmo if we don't intend on resizing it. Close it. 98353944Sdim if (Flags & MAP_RESIZABLE) { 99353944Sdim DCHECK(Data); 100353944Sdim DCHECK_EQ(Data->Vmo, ZX_HANDLE_INVALID); 101353944Sdim Data->Vmo = Vmo; 102353944Sdim } else { 103353944Sdim CHECK_EQ(_zx_handle_close(Vmo), ZX_OK); 104353944Sdim } 105353944Sdim if (UNLIKELY(Status != ZX_OK)) { 106353944Sdim if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) 107353944Sdim dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY); 108353944Sdim return nullptr; 109353944Sdim } 110353944Sdim if (Data) 111353944Sdim Data->VmoSize += Size; 112353944Sdim 113353944Sdim return reinterpret_cast<void *>(P); 114353944Sdim} 115353944Sdim 116353944Sdimvoid unmap(void *Addr, uptr Size, uptr Flags, MapPlatformData *Data) { 117353944Sdim if (Flags & UNMAP_ALL) { 118353944Sdim DCHECK_NE(Data, nullptr); 119353944Sdim const zx_handle_t Vmar = Data->Vmar; 120353944Sdim DCHECK_NE(Vmar, _zx_vmar_root_self()); 121353944Sdim // Destroying the vmar effectively unmaps the whole mapping. 122353944Sdim CHECK_EQ(_zx_vmar_destroy(Vmar), ZX_OK); 123353944Sdim CHECK_EQ(_zx_handle_close(Vmar), ZX_OK); 124353944Sdim } else { 125353944Sdim const zx_handle_t Vmar = Data ? Data->Vmar : _zx_vmar_root_self(); 126353944Sdim const zx_status_t Status = 127353944Sdim _zx_vmar_unmap(Vmar, reinterpret_cast<uintptr_t>(Addr), Size); 128353944Sdim if (UNLIKELY(Status != ZX_OK)) 129353944Sdim dieOnMapUnmapError(); 130353944Sdim } 131353944Sdim if (Data) { 132353944Sdim if (Data->Vmo != ZX_HANDLE_INVALID) 133353944Sdim CHECK_EQ(_zx_handle_close(Data->Vmo), ZX_OK); 134353944Sdim memset(Data, 0, sizeof(*Data)); 135353944Sdim } 136353944Sdim} 137353944Sdim 138353944Sdimvoid releasePagesToOS(UNUSED uptr BaseAddress, uptr Offset, uptr Size, 139353944Sdim MapPlatformData *Data) { 140353944Sdim DCHECK(Data); 141353944Sdim DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID); 142353944Sdim DCHECK_NE(Data->Vmo, ZX_HANDLE_INVALID); 143353944Sdim const zx_status_t Status = 144353944Sdim _zx_vmo_op_range(Data->Vmo, ZX_VMO_OP_DECOMMIT, Offset, Size, NULL, 0); 145353944Sdim CHECK_EQ(Status, ZX_OK); 146353944Sdim} 147353944Sdim 148353944Sdimconst char *getEnv(const char *Name) { return getenv(Name); } 149353944Sdim 150353944Sdim// Note: we need to flag these methods with __TA_NO_THREAD_SAFETY_ANALYSIS 151353944Sdim// because the Fuchsia implementation of sync_mutex_t has clang thread safety 152353944Sdim// annotations. Were we to apply proper capability annotations to the top level 153353944Sdim// HybridMutex class itself, they would not be needed. As it stands, the 154353944Sdim// thread analysis thinks that we are locking the mutex and accidentally leaving 155353944Sdim// it locked on the way out. 156353944Sdimbool HybridMutex::tryLock() __TA_NO_THREAD_SAFETY_ANALYSIS { 157353944Sdim // Size and alignment must be compatible between both types. 158353944Sdim return sync_mutex_trylock(&M) == ZX_OK; 159353944Sdim} 160353944Sdim 161353944Sdimvoid HybridMutex::lockSlow() __TA_NO_THREAD_SAFETY_ANALYSIS { 162353944Sdim sync_mutex_lock(&M); 163353944Sdim} 164353944Sdim 165353944Sdimvoid HybridMutex::unlock() __TA_NO_THREAD_SAFETY_ANALYSIS { 166353944Sdim sync_mutex_unlock(&M); 167353944Sdim} 168353944Sdim 169353944Sdimu64 getMonotonicTime() { return _zx_clock_get_monotonic(); } 170353944Sdim 171353944Sdimu32 getNumberOfCPUs() { return _zx_system_get_num_cpus(); } 172353944Sdim 173353944Sdimbool getRandom(void *Buffer, uptr Length, UNUSED bool Blocking) { 174357095Sdim static_assert(MaxRandomLength <= ZX_CPRNG_DRAW_MAX_LEN, ""); 175353944Sdim if (UNLIKELY(!Buffer || !Length || Length > MaxRandomLength)) 176353944Sdim return false; 177353944Sdim _zx_cprng_draw(Buffer, Length); 178353944Sdim return true; 179353944Sdim} 180353944Sdim 181353944Sdimvoid outputRaw(const char *Buffer) { 182353944Sdim __sanitizer_log_write(Buffer, strlen(Buffer)); 183353944Sdim} 184353944Sdim 185353944Sdimvoid setAbortMessage(const char *Message) {} 186353944Sdim 187353944Sdim} // namespace scudo 188353944Sdim 189353944Sdim#endif // SCUDO_FUCHSIA 190