/fuchsia/zircon/third_party/ulib/backtrace/ |
H A D | mmapio.c | 54 int descriptor, off_t offset, size_t size, 67 size += inpage; 68 size = (size + (pagesize - 1)) & ~ (pagesize - 1); 70 map = mmap (NULL, size, PROT_READ, MAP_PRIVATE, descriptor, pageoff); 79 view->len = size; 53 backtrace_get_view(struct backtrace_state *state ATTRIBUTE_UNUSED, int descriptor, off_t offset, size_t size, backtrace_error_callback error_callback, void *data, struct backtrace_view *view) argument
|
H A D | read.c | 49 off_t offset, size_t size, 61 view->base = backtrace_alloc (state, size, error_callback, data); 65 view->len = size; 67 got = read (descriptor, view->base, size); 75 if ((size_t) got < size) 48 backtrace_get_view(struct backtrace_state *state, int descriptor, off_t offset, size_t size, backtrace_error_callback error_callback, void *data, struct backtrace_view *view) argument
|
/fuchsia/zircon/third_party/ulib/musl/src/passwd/ |
H A D | getgr_r.c | 5 static int getgr_r(const char* name, gid_t gid, struct group* gr, char* buf, size_t size, argument 15 if (*res && size < len + (nmem + 1) * sizeof(char*) + 32) { 35 int getgrnam_r(const char* name, struct group* gr, char* buf, size_t size, struct group** res) { argument 36 return getgr_r(name, 0, gr, buf, size, res); 39 int getgrgid_r(gid_t gid, struct group* gr, char* buf, size_t size, struct group** res) { argument 40 return getgr_r(0, gid, gr, buf, size, res);
|
H A D | getpw_r.c | 5 static int getpw_r(const char* name, uid_t uid, struct passwd* pw, char* buf, size_t size, argument 12 if (*res && size < len) { 28 int getpwnam_r(const char* name, struct passwd* pw, char* buf, size_t size, struct passwd** res) { argument 29 return getpw_r(name, 0, pw, buf, size, res); 32 int getpwuid_r(uid_t uid, struct passwd* pw, char* buf, size_t size, struct passwd** res) { argument 33 return getpw_r(0, uid, pw, buf, size, res);
|
/fuchsia/zircon/third_party/ulib/jemalloc/src/ |
H A D | base.c | 12 base_map(extent_hooks_t *extent_hooks, unsigned ind, size_t size) argument 18 assert(size == HUGEPAGE_CEILING(size)); 21 addr = extent_alloc_mmap(NULL, size, PAGE, &zero, &commit); 23 addr = extent_hooks->alloc(extent_hooks, NULL, size, PAGE, 31 base_unmap(extent_hooks_t *extent_hooks, unsigned ind, void *addr, size_t size) argument 44 if (!extent_dalloc_mmap(addr, size)) 46 if (!pages_decommit(addr, size)) 48 if (!pages_purge_lazy(addr, size)) 50 if (!pages_purge_forced(addr, size)) 75 base_extent_init(size_t *extent_sn_next, extent_t *extent, void *addr, size_t size) argument 87 base_extent_bump_alloc_helper(extent_t *extent, size_t *gap_size, size_t size, size_t alignment) argument 106 base_extent_bump_alloc_post(tsdn_t *tsdn, base_t *base, extent_t *extent, size_t gap_size, void *addr, size_t size) argument 133 base_extent_bump_alloc(tsdn_t *tsdn, base_t *base, extent_t *extent, size_t size, size_t alignment) argument 150 base_block_alloc(extent_hooks_t *extent_hooks, unsigned ind, size_t *extent_sn_next, size_t size, size_t alignment) argument 177 base_extent_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) argument 290 base_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) argument [all...] |
H A D | zone.c | 13 size_t (*size)(struct _malloc_zone_t *, const void *); member in struct:_malloc_zone_t 32 vm_size_t size; member in struct:__anon1585 93 static void *zone_malloc(malloc_zone_t *zone, size_t size); 94 static void *zone_calloc(malloc_zone_t *zone, size_t num, size_t size); 95 static void *zone_valloc(malloc_zone_t *zone, size_t size); 97 static void *zone_realloc(malloc_zone_t *zone, void *ptr, size_t size); 99 size_t size); 101 size_t size); 103 static unsigned zone_batch_malloc(struct _malloc_zone_t *zone, size_t size, 108 static size_t zone_good_size(malloc_zone_t *zone, size_t size); 143 zone_malloc(malloc_zone_t *zone, size_t size) argument 149 zone_calloc(malloc_zone_t *zone, size_t num, size_t size) argument 155 zone_valloc(malloc_zone_t *zone, size_t size) argument 176 zone_realloc(malloc_zone_t *zone, void *ptr, size_t size) argument 185 zone_memalign(malloc_zone_t *zone, size_t alignment, size_t size) argument 195 zone_free_definite_size(malloc_zone_t *zone, void *ptr, size_t size) argument 217 zone_batch_malloc(struct _malloc_zone_t *zone, size_t size, void **results, unsigned num_requested) argument 250 zone_good_size(malloc_zone_t *zone, size_t size) argument [all...] |
/fuchsia/zircon/kernel/vm/ |
H A D | vm_address_region.cpp | 24 VmAddressRegion::VmAddressRegion(VmAspace& aspace, vaddr_t base, size_t size, uint32_t vmar_flags) argument 25 : VmAddressRegionOrMapping(base, size, vmar_flags | VMAR_CAN_RWX_FLAGS, 35 VmAddressRegion::VmAddressRegion(VmAddressRegion& parent, vaddr_t base, size_t size, argument 37 : VmAddressRegionOrMapping(base, size, vmar_flags, parent.aspace_.get(), 45 : VmAddressRegion(kernel_aspace, kernel_aspace.base(), kernel_aspace.size(), 64 auto vmar = new (&ac) VmAddressRegion(aspace, aspace.base(), aspace.size(), vmar_flags); 74 zx_status_t VmAddressRegion::CreateSubVmarInternal(size_t offset, size_t size, uint8_t align_pow2, argument 86 if (size == 0) { 129 if (offset >= size_ || size > size_ - offset) { 142 if (!IsRangeAvailableLocked(new_base, size)) { 193 CreateSubVmar(size_t offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, const char* name, fbl::RefPtr<VmAddressRegion>* out) argument 218 CreateVmMapping(size_t mapping_offset, size_t size, uint8_t align_pow2, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, const char* name, fbl::RefPtr<VmMapping>* out) argument 269 OverwriteVmMapping( vaddr_t base, size_t size, uint32_t vmar_flags, fbl::RefPtr<VmObject> vmo, uint64_t vmo_offset, uint arch_mmu_flags, fbl::RefPtr<VmAddressRegionOrMapping>* out) argument 401 IsRangeAvailableLocked(vaddr_t base, size_t size) argument 518 AllocSpotLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 615 Unmap(vaddr_t base, size_t size) argument 632 UnmapAllowPartial(vaddr_t base, size_t size) argument 662 UnmapInternalLocked(vaddr_t base, size_t size, bool can_destroy_regions, bool allow_partial_vmar) argument 775 Protect(vaddr_t base, size_t size, uint new_arch_mmu_flags) argument 855 LinearRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 927 NonCompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument 998 CompactRandomizedRegionAllocatorLocked(size_t size, uint8_t align_pow2, uint arch_mmu_flags, vaddr_t* spot) argument [all...] |
/fuchsia/zircon/system/dev/bus/platform/ |
H A D | device-resources.h | 49 inline size_t mmio_count() const { return mmios_.size(); } 50 inline size_t irq_count() const { return irqs_.size(); } 51 inline size_t gpio_count() const { return gpios_.size(); } 52 inline size_t i2c_channel_count() const { return i2c_channels_.size(); } 53 inline size_t clk_count() const { return clks_.size(); } 54 inline size_t bti_count() const { return btis_.size(); } 55 inline size_t metadata_count() const { return metadata_.size(); } 56 inline size_t boot_metadata_count() const { return boot_metadata_.size(); } 57 inline size_t child_count() const { return children_.size(); } 58 inline size_t protocol_count() const { return protocols_.size(); } [all...] |
/fuchsia/zircon/kernel/dev/iommu/dummy/ |
H A D | dummy_iommu.cpp | 42 uint64_t offset, size_t size, uint32_t perms, 47 if (!IS_PAGE_ALIGNED(offset) || size == 0) { 56 if (offset + size < offset || offset + size > vmo->size()) { 67 zx_status_t status = vmo->Lookup(offset, fbl::min<size_t>(PAGE_SIZE, size), 0, lookup_fn, 81 *mapped_len = ROUNDUP(size, PAGE_SIZE); 87 uint64_t offset, size_t size, uint32_t perms, 92 if (!IS_PAGE_ALIGNED(offset) || size == 0) { 102 if (add_overflow(offset, size, 41 Map(uint64_t bus_txn_id, const fbl::RefPtr<VmObject>& vmo, uint64_t offset, size_t size, uint32_t perms, dev_vaddr_t* vaddr, size_t* mapped_len) argument 86 MapContiguous(uint64_t bus_txn_id, const fbl::RefPtr<VmObject>& vmo, uint64_t offset, size_t size, uint32_t perms, dev_vaddr_t* vaddr, size_t* mapped_len) argument 130 Unmap(uint64_t bus_txn_id, dev_vaddr_t vaddr, size_t size) argument [all...] |
/fuchsia/zircon/kernel/lib/crypto/ |
H A D | prng.cpp | 30 PRNG::PRNG(const void* data, size_t size) argument 31 : PRNG(data, size, NonThreadSafeTag()) { 35 PRNG::PRNG(const void* data, size_t size, NonThreadSafeTag tag) argument 39 AddEntropy(data, size); 47 void PRNG::AddEntropy(const void* data, size_t size) { argument 48 DEBUG_ASSERT(data || size == 0); 49 ASSERT(size <= kMaxEntropy); 60 SHA256_Update(&ctx, data, size); 73 accumulated_.fetch_add(size) + size > 78 Draw(void* out, size_t size) argument [all...] |
/fuchsia/zircon/system/ulib/bitmap/include/bitmap/ |
H A D | storage.h | 30 zx_status_t Allocate(size_t size) { argument 32 auto arr = new (&ac) uint8_t[size]; 36 storage_.reset(arr, size); 51 zx_status_t Allocate(size_t size) { argument 52 ZX_ASSERT(size <= N); 74 zx_status_t Allocate(size_t size) { argument 76 size_ = fbl::round_up(size, static_cast<size_t>(PAGE_SIZE)); 90 zx_status_t Grow(size_t size) { argument 91 if (size <= size_) { 95 size [all...] |
/fuchsia/zircon/third_party/ulib/musl/include/ |
H A D | sched.h | 56 #define __CPU_op_S(i, size, set, op) \ 57 ((i) / 8U >= (size) \ 61 #define CPU_SET_S(i, size, set) __CPU_op_S(i, size, set, |=) 62 #define CPU_CLR_S(i, size, set) __CPU_op_S(i, size, set, &= ~) 63 #define CPU_ISSET_S(i, size, set) __CPU_op_S(i, size, set, &) 79 #define CPU_COUNT_S(size, set) __sched_cpucount(size, se [all...] |
/fuchsia/zircon/third_party/ulib/musl/src/stdio/ |
H A D | fmemopen.c | 7 size_t pos, len, size; member in struct:cookie 21 if (off < -base || off > (ssize_t)c->size - base) 58 rem = c->size - c->pos; 65 if (c->len < c->size) 67 else if ((f->flags & F_NORD) && c->size) 68 c->buf[c->size - 1] = 0; 77 FILE* fmemopen(void* restrict buf, size_t size, const char* restrict mode) { argument 82 if (!size || !strchr("rwa", *mode)) { 87 if (!buf && size > SIZE_MAX - sizeof(FILE) - BUFSIZ - UNGET) { 92 f = calloc(sizeof *f + sizeof *c + UNGET + BUFSIZ + (buf ? 0 : size), [all...] |
H A D | vfwscanf.c | 23 static void store_int(void* dest, int size, unsigned long long i) { argument 26 switch (size) { 90 int size; local 156 size = SIZE_def; 160 p++, size = SIZE_hh; 162 size = SIZE_h; 166 p++, size = SIZE_ll; 168 size = SIZE_l; 171 size = SIZE_ll; 175 size [all...] |
/fuchsia/zircon/kernel/lib/unittest/ |
H A D | user_memory.cpp | 15 zx_status_t status = mapping_->Unmap(mapping_->base(), mapping_->size()); 20 fbl::unique_ptr<UserMemory> UserMemory::Create(size_t size) { argument 21 size = ROUNDUP_PAGE_SIZE(size); 24 zx_status_t status = VmObjectPaged::Create(PMM_ALLOC_FLAG_ANY, 0u, size, &vmo); 40 status = root_vmar->CreateVmMapping(/* offset= */ 0, size, /* align_pow2= */ 0, vmar_flags, 48 zx_status_t status = mapping->Unmap(mapping->base(), mapping->size());
|
/fuchsia/zircon/scripts/ |
H A D | gen-rodso-code.sh | 41 local symbol type addr size rest 42 while read symbol type addr size rest; do 49 case "$size" in 51 *) echo "#define ${1}_${symbol}_SIZE 0x${size}" >> $OUTFILE 106 # ^^^^^ addr ^^^^^ ^size^ ^entsize^ 111 local size=$(printf '%#x' 0x${words[2]}) 114 if [ $size != $entsize ]; then 117 echo "#define ${1}_DATA_END_dynsym (${1}_DATA_START_dynsym + $size)" >> $OUTFILE
|
/fuchsia/zircon/system/ulib/zx/include/lib/zx/ |
H A D | vmo.h | 29 static zx_status_t create(uint64_t size, uint32_t options, vmo* result); 39 zx_status_t get_size(uint64_t* size) const { 40 return zx_vmo_get_size(get(), size); 43 zx_status_t set_size(uint64_t size) const { 44 return zx_vmo_set_size(get(), size); 47 zx_status_t clone(uint32_t options, uint64_t offset, uint64_t size, argument 52 get(), options, offset, size, h.reset_and_get_address()); 57 zx_status_t op_range(uint32_t op, uint64_t offset, uint64_t size, argument 59 return zx_vmo_op_range(get(), op, offset, size, buffer, buffer_size);
|
/fuchsia/zircon/system/utest/vmo/ |
H A D | vmo.cpp | 306 // get the size that we set it to 307 uint64_t size = 0x99999999; local 308 status = zx_vmo_get_size(vmo, &size); 310 EXPECT_EQ(len, size, "vm_object_get_size"); 317 // get the size again 318 size = 0x99999999; 319 status = zx_vmo_get_size(vmo, &size); 321 EXPECT_EQ(len, size, "vm_object_get_size"); 323 // try to resize it to a ludicrous size 327 // resize it to a non aligned size 380 size_t size; local 478 uint64_t size = 0x99999999; local 481 EXPECT_EQ(fbl::round_up(s, static_cast<size_t>(PAGE_SIZE)), size, "vm_object_get_size"); local 505 uint64_t size = 0x99999999; local 508 EXPECT_EQ(fbl::round_up(s, static_cast<size_t>(PAGE_SIZE)), size, "vm_object_get_size"); local 530 uint64_t size = 0x99999999; local 533 EXPECT_EQ(fbl::round_up(s, static_cast<size_t>(PAGE_SIZE)), size, "vm_object_get_size"); local 725 const size_t size = 16384; local 816 const size_t size = PAGE_SIZE * 4; local 891 const size_t size = PAGE_SIZE * 4; local 931 const size_t size = PAGE_SIZE * 4; local 1015 const size_t size = PAGE_SIZE * 4; local 1094 const size_t size = PAGE_SIZE * 4; local 1161 const size_t size = PAGE_SIZE * 4; local 1221 const size_t size = PAGE_SIZE; local 1338 const size_t size = 0x8000; local 1375 const size_t size = 0x8000; local 1403 zx_vmar_unmap(zx_vmar_root_self(), ptr_rw, size); local 1404 zx_vmar_unmap(zx_vmar_root_self(), ptr_ro, size); local 1440 const size_t size = PAGE_SIZE * 4; local 1585 const size_t size = PAGE_SIZE * 2; local 1612 const size_t size = PAGE_SIZE * 2; local 1643 const size_t size = PAGE_SIZE * 2; local [all...] |
/fuchsia/zircon/kernel/target/arm64/boot-shim/ |
H A D | devicetree.c | 28 if (off >= src->size) 30 if (len >= src->size) 32 if ((off + len) > src->size) 35 dst->size = len; 41 return s->size == 0; 48 if (s->size < 4) { 49 s->size = 0; 53 s->size -= 4; 63 if (len > s->size) { 64 s->size [all...] |
H A D | devicetree.h | 14 uint32_t size; member in struct:dt_slice 19 uint32_t size; member in struct:devicetree_header 26 uint32_t sz_strings; // size of stringdata 27 uint32_t sz_struct; // size of DT 'structure' 39 typedef int (*dt_prop_cb)(const char *name, uint8_t *data, uint32_t size, void *cookie);
|
/fuchsia/zircon/system/utest/libfzl/ |
H A D | vmo-vmar-tests.cpp | 29 uintptr_t contained_end = contained_start + contained.size(); 31 uintptr_t container_end = container_start + container.size(); 164 vmo_end += mappers[i].size(); 165 vmar_end += target_vmar->size(); 196 // If we mapped this VMO, and we passed zero for the map size, the Mapper should 197 // have mapped the entire VMO after the offset and its size should reflect that. 199 ASSERT_EQ(mappers[i].size() + t.test_offset, kVmoTestSize); 295 size_t size; local 301 ASSERT_EQ(mapper1.size(), 0); 303 ASSERT_EQ(mapper2.size(), [all...] |
/fuchsia/zircon/kernel/dev/iommu/intel/ |
H A D | device_context.cpp | 88 uint64_t size = aspace_size() - base; local 89 region_alloc_.AddRegion({ .base = 1ull << 20, .size = size }); 196 size_t size, uint32_t perms, bool map_contiguous, 203 return SecondLevelMapDiscontiguous(vmo, offset, size, flags, map_contiguous, 206 return SecondLevelMapContiguous(vmo, offset, size, flags, virt_paddr, mapped_len); 210 uint64_t offset, size_t size, uint flags, 216 if (!map_contiguous && size > min_contig) { 217 size = min_contig; 227 zx_status_t status = region_alloc_.GetRegion(size, min_conti 195 SecondLevelMap(const fbl::RefPtr<VmObject>& vmo, uint64_t offset, size_t size, uint32_t perms, bool map_contiguous, paddr_t* virt_paddr, size_t* mapped_len) argument 209 SecondLevelMapDiscontiguous(const fbl::RefPtr<VmObject>& vmo, uint64_t offset, size_t size, uint flags, bool map_contiguous, paddr_t* virt_paddr, size_t* mapped_len) argument 286 SecondLevelMapContiguous(const fbl::RefPtr<VmObject>& vmo, uint64_t offset, size_t size, uint flags, paddr_t* virt_paddr, size_t* mapped_len) argument 343 SecondLevelMapIdentity(paddr_t base, size_t size, uint32_t perms) argument 376 SecondLevelUnmap(paddr_t virt_paddr, size_t size) argument [all...] |
/fuchsia/zircon/kernel/tests/ |
H A D | resource_tests.cpp | 62 uint64_t size = PAGE_SIZE; local 68 EXPECT_EQ(ResourceDispatcher::Create(&disp1, &rights, ZX_RSRC_KIND_MMIO, base, size, 75 EXPECT_EQ(ResourceDispatcher::Create(&disp2, &rights, ZX_RSRC_KIND_MMIO, base, size, 93 uint64_t size = PAGE_SIZE; local 99 EXPECT_EQ(ResourceDispatcher::Create(&disp1, &rights, ZX_RSRC_KIND_MMIO, base, size, 106 EXPECT_EQ(ResourceDispatcher::Create(&disp2, &rights, ZX_RSRC_KIND_MMIO, base, size, 122 uint64_t size = 0xFFFF; local 124 ASSERT_EQ(ResourceDispatcher::InitializeAllocator(ZX_RSRC_KIND_MMIO, 0, size, test_rallocs), 127 EXPECT_EQ(ResourceDispatcher::Create(&disp1, &rights, ZX_RSRC_KIND_MMIO, size - 0xFF, 0xFFF, 132 EXPECT_EQ(ResourceDispatcher::Create(&disp1, &rights, ZX_RSRC_KIND_MMIO, size [all...] |
/fuchsia/zircon/system/dev/pci/amlogic-pcie/ |
H A D | aml-pcie.h | 41 void SetMaxPayload(const uint32_t size); 42 void SetMaxReadRequest(const uint32_t size); 46 void RmwCtrlSts(const uint32_t size, const uint32_t shift, const uint32_t mask);
|
/fuchsia/zircon/system/private/zircon/ |
H A D | initializer_list.h | 35 constexpr size_t size() const { return size_; } function in class:std::initializer_list 40 constexpr initializer_list(const T* items, size_t size) argument 41 : items_(items), size_(size) {}
|