/freebsd-current/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_allocator_secondary.h | 85 void *Allocate(AllocatorStats *stat, const uptr size, uptr alignment) { argument 86 CHECK(IsPowerOfTwo(alignment)); 88 if (alignment > page_size_) 89 map_size += alignment; 93 "0x%zx bytes with 0x%zx alignment requested\n", 94 SanitizerToolName, map_size, alignment); 104 if (res & (alignment - 1)) // Align. 105 res += alignment - (res & (alignment - 1)); 107 CHECK(IsAligned(res, alignment)); [all...] |
/freebsd-current/contrib/llvm-project/compiler-rt/lib/asan/ |
H A D | asan_malloc_linux.cpp | 141 INTERCEPTOR(int, posix_memalign, void **memptr, uptr alignment, uptr size) { argument 143 return asan_posix_memalign(memptr, alignment, size, &stack); 171 void *(*memalign)(uptr alignment, uptr bytes); 181 void *(*memalign)(uptr alignment, uptr bytes); 182 int (*posix_memalign)(void **memptr, uptr alignment, uptr size);
|
H A D | asan_allocator.cpp | 528 void *Allocate(uptr size, uptr alignment, BufferedStackTrace *stack, argument 541 ComputeUserRequestedAlignmentLog(alignment); 542 if (alignment < min_alignment) 543 alignment = min_alignment; 553 CHECK(IsPowerOfTwo(alignment)); 556 uptr rounded_size = RoundUpTo(Max(size, kChunkHeader2Size), alignment); 558 if (alignment > min_alignment) 559 needed_size += alignment; 560 bool from_primary = PrimaryAllocator::CanAllocate(needed_size, alignment); 598 if (!IsAligned(user_beg, alignment)) 994 asan_delete(void *ptr, uptr size, uptr alignment, BufferedStackTrace *stack, AllocType alloc_type) argument 1051 asan_memalign(uptr alignment, uptr size, BufferedStackTrace *stack, AllocType alloc_type) argument 1063 asan_aligned_alloc(uptr alignment, uptr size, BufferedStackTrace *stack) argument 1074 asan_posix_memalign(void **memptr, uptr alignment, uptr size, BufferedStackTrace *stack) argument [all...] |
H A D | asan_allocator.h | 61 u32 UserRequestedAlignment() const; // Originally requested alignment. 270 void *asan_memalign(uptr alignment, uptr size, BufferedStackTrace *stack, 273 void asan_delete(void *ptr, uptr size, uptr alignment, 284 void *asan_aligned_alloc(uptr alignment, uptr size, BufferedStackTrace *stack); 285 int asan_posix_memalign(void **memptr, uptr alignment, uptr size,
|
/freebsd-current/sys/x86/x86/ |
H A D | busdma_bounce.c | 110 #define dmat_alignment(dmat) ((dmat)->common.alignment) 171 bounce_bus_dma_tag_create(bus_dma_tag_t parent, bus_size_t alignment, argument 181 NULL, alignment, boundary, lowaddr, highaddr, maxsize, nsegments, 204 newtag->common.alignment > 1) 306 * exclusion region, a data alignment that is stricter than 1, and/or 320 if (dmat->common.alignment > 1) 421 * - Its alignment requirement is also smaller than the page size. 428 * - The alignment constraint isn't larger than a page boundary. 436 dmat->common.alignment <= PAGE_SIZE && 440 dmat->common.alignment, M_DEVBU [all...] |
/freebsd-current/contrib/tcpdump/ |
H A D | cpack.c | 44 nd_cpack_next_boundary(const uint8_t *buf, const uint8_t *p, size_t alignment) argument 46 size_t misalignment = (size_t)(p - buf) % alignment; 51 return p + (alignment - misalignment); 63 /* Ensure alignment. */
|
/freebsd-current/contrib/llvm-project/compiler-rt/lib/hwasan/ |
H A D | hwasan.h | 73 // Minimal alignment of the shadow base address. Determines the space available 133 void *hwasan_aligned_alloc(uptr alignment, uptr size, StackTrace *stack); 134 void *hwasan_memalign(uptr alignment, uptr size, StackTrace *stack); 135 int hwasan_posix_memalign(void **memptr, uptr alignment, uptr size,
|
/freebsd-current/lib/geom/part/ |
H A D | geom_part.c | 96 { 'a', "alignment", GPART_AUTOFILL, G_TYPE_STRING }, 104 "-t type [-a alignment] [-b start] [-s size] [-i index] " 174 { 'a', "alignment", GPART_AUTOFILL, G_TYPE_STRING }, 179 "-i index [-a alignment] [-s size] [-f flags] geom" 327 off_t lba, new_lba, alignment, offset; local 354 s = gctl_get_ascii(req, "alignment"); 356 alignment = 1; 358 error = g_parse_lba(s, pp->lg_sectorsize, &alignment); 360 errc(EXIT_FAILURE, error, "Invalid alignment param"); 361 if (alignment 451 off_t lba, len, alignment, offset; local [all...] |
/freebsd-current/sys/dev/pms/RefTisa/tisa/sassata/sas/ini/ |
H A D | itdinit.c | 130 iniMem->tdCachedMem[0].alignment = sizeof (void *); /* 4 bytes */ 147 iniMem->tdCachedMem[i].alignment = 0; 170 initiatorResource->initiatorOption.dynamicDmaMem.alignment = 0; 176 initiatorResource->initiatorOption.dynamicCachedMem.alignment = 0; 197 initiatorResource->initiatorOption.dynamicDmaMem.alignment = sizeof(void *); 214 initiatorResource->initiatorOption.dynamicCachedMem.alignment = sizeof(void *);
|
/freebsd-current/contrib/jemalloc/src/ |
H A D | large.c | 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument 30 ausize = sz_sa2u(usize, alignment); 48 arena, usize, alignment, &is_zeroed)) == NULL) { 266 size_t alignment, bool zero) { 267 if (alignment <= CACHELINE) { 270 return large_palloc(tsdn, arena, usize, alignment, zero); 275 size_t alignment, bool zero, tcache_t *tcache, 299 void *ret = large_ralloc_move_helper(tsdn, arena, usize, alignment, 265 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 274 large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize, size_t alignment, bool zero, tcache_t *tcache, hook_ralloc_args_t *hook_args) argument
|
H A D | extent.c | 26 size_t size, size_t alignment, bool *zero, bool *commit, 105 size_t usize, size_t pad, size_t alignment, bool slab, szind_t szind, 402 * Find an extent with size [min_size, max_size) to satisfy the alignment 407 size_t alignment) { 423 PAGE_CEILING(alignment)); 425 /* Overflow or not crossing the next alignment. */ 494 size_t esize, size_t alignment) { 497 size_t max_size = esize + PAGE_CEILING(alignment) - PAGE; 506 if (alignment > PAGE && extent == NULL) { 508 * max_size guarantees the alignment requiremen 406 extents_fit_alignment(extents_t *extents, size_t min_size, size_t max_size, size_t alignment) argument 493 extents_fit_locked(tsdn_t *tsdn, arena_t *arena, extents_t *extents, size_t esize, size_t alignment) argument 537 extents_alloc(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, extents_t *extents, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit) argument 892 extent_recycle_extract(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, rtree_ctx_t *rtree_ctx, extents_t *extents, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, bool growing_retained) argument 980 extent_split_interior(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, rtree_ctx_t *rtree_ctx, extent_t **extent, extent_t **lead, extent_t **trail, extent_t **to_leak, extent_t **to_salvage, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool growing_retained) argument 1057 extent_recycle_split(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, rtree_ctx_t *rtree_ctx, extents_t *extents, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, extent_t *extent, bool growing_retained) argument 1128 extent_recycle(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, extents_t *extents, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit, bool growing_retained) argument 1209 extent_alloc_core(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, dss_prec_t dss_prec) argument 1239 extent_alloc_default_impl(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument 1251 extent_alloc_default(extent_hooks_t *extent_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, unsigned arena_ind) argument 1297 extent_grow_retained(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit) argument 1463 extent_alloc_retained(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit) argument 1492 extent_alloc_wrapper_hard(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit) argument 1531 extent_alloc_wrapper(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks, void *new_addr, size_t size, size_t pad, size_t alignment, bool slab, szind_t szind, bool *zero, bool *commit) argument [all...] |
H A D | jemalloc.c | 1825 * Whether or not allocations (with alignment) of size 0 should be 1844 * The minimum valid alignment for functions requesting aligned storage. 1850 /* The error string to use if the passed-in alignment is invalid. */ 1896 size_t alignment; member in struct:dynamic_opts_s 1908 dynamic_opts->alignment = 0; 1914 /* ind is ignored if dopts->alignment > 0. */ 1948 if (unlikely(dopts->alignment != 0)) { 1949 return ipalloct(tsd_tsdn(tsd), usize, dopts->alignment, 1970 assert(((dopts->alignment == 0) ? 1972 sz_sa2u(SC_LARGE_MINCLASS, dopts->alignment)) 2397 je_posix_memalign(void **memptr, size_t alignment, size_t size) argument 3134 irallocx_prof_sample(tsdn_t *tsdn, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx, hook_ralloc_args_t *hook_args) argument 3159 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena, alloc_ctx_t *alloc_ctx, hook_ralloc_args_t *hook_args) argument 3206 size_t alignment = MALLOCX_ALIGN_GET(flags); local 3297 ixallocx_helper(tsdn_t *tsdn, void *ptr, size_t old_usize, size_t size, size_t extra, size_t alignment, bool zero) argument 3310 ixallocx_prof_sample(tsdn_t *tsdn, void *ptr, size_t old_usize, size_t size, size_t extra, size_t alignment, bool zero, prof_tctx_t *tctx) argument 3324 ixallocx_prof(tsd_t *tsd, void *ptr, size_t old_usize, size_t size, size_t extra, size_t alignment, bool zero, alloc_ctx_t *alloc_ctx) argument 3378 size_t alignment = MALLOCX_ALIGN_GET(flags); local [all...] |
/freebsd-current/contrib/llvm-project/compiler-rt/lib/memprof/ |
H A D | memprof_malloc_linux.cpp | 130 INTERCEPTOR(int, posix_memalign, void **memptr, uptr alignment, uptr size) { argument 132 return memprof_posix_memalign(memptr, alignment, size, &stack);
|
/freebsd-current/usr.sbin/bhyve/ |
H A D | basl.h | 86 const uint8_t *fwcfg_name, uint32_t alignment, 99 const uint8_t *name, uint32_t alignment);
|
/freebsd-current/sys/dev/qat/qat_common/ |
H A D | qat_freebsd.c | 42 bus_size_t alignment, 52 alignment, 40 bus_dma_mem_create(struct bus_dmamem *mem, bus_dma_tag_t parent, bus_size_t alignment, bus_addr_t lowaddr, bus_size_t len, int flags) argument
|
/freebsd-current/contrib/llvm-project/lldb/bindings/interface/ |
H A D | SBSectionExtensions.i | 31 alignment = property(GetAlignment, None, doc='''A read only property that returns the alignment of this section as a number of host bytes.''')
|
/freebsd-current/contrib/llvm-project/compiler-rt/lib/lsan/ |
H A D | lsan_interceptors.cpp | 108 INTERCEPTOR(int, posix_memalign, void **memptr, uptr alignment, uptr size) { argument 111 return lsan_posix_memalign(memptr, alignment, size, stack); 122 INTERCEPTOR(void*, memalign, uptr alignment, uptr size) { argument 125 return lsan_memalign(alignment, size, stack); 133 INTERCEPTOR(void *, __libc_memalign, uptr alignment, uptr size) { argument 136 void *res = lsan_memalign(alignment, size, stack); 146 INTERCEPTOR(void*, aligned_alloc, uptr alignment, uptr size) { argument 149 return lsan_aligned_alloc(alignment, size, stack);
|
/freebsd-current/lib/libomp/ |
H A D | Makefile | 69 CWARNFLAGS+= -Wno-atomic-alignment 73 # builtin operations must have natural alignment, but there is not much we 75 CWARNFLAGS+= -Wno-sync-alignment
|
/freebsd-current/contrib/jemalloc/include/jemalloc/internal/ |
H A D | pages.h | 75 void *pages_map(void *addr, size_t size, size_t alignment, bool *commit);
|
/freebsd-current/sys/dev/pms/RefTisa/discovery/api/ |
H A D | dm.h | 127 bit32 alignment; member in struct:__anon4152
|
/freebsd-current/sys/powerpc/pseries/ |
H A D | plpar_iommu.c | 160 bus_addr_t min, bus_addr_t max, bus_size_t alignment, bus_addr_t boundary, 181 (alignment < PAGE_SIZE) ? PAGE_SIZE : alignment, 0, 189 alignment == 0, 190 ("Allocated segment does not match alignment constraint")); 159 phyp_iommu_map(device_t dev, bus_dma_segment_t *segs, int *nsegs, bus_addr_t min, bus_addr_t max, bus_size_t alignment, bus_addr_t boundary, void *cookie) argument
|
/freebsd-current/sys/arm/arm/ |
H A D | bus_space_generic.c | 84 bus_size_t alignment, bus_size_t boundary, int flags, bus_addr_t *bpap, 83 generic_bs_alloc(bus_space_tag_t t, bus_addr_t rstart, bus_addr_t rend, bus_size_t size, bus_size_t alignment, bus_size_t boundary, int flags, bus_addr_t *bpap, bus_space_handle_t *bshp) argument
|
/freebsd-current/contrib/llvm-project/lld/ELF/ |
H A D | Thunks.h | 63 // The alignment requirement for this Thunk, defaults to the size of the 64 // typical code section alignment. 65 uint32_t alignment = 4; member in class:lld::elf::Thunk
|
/freebsd-current/contrib/llvm-project/clang/include/clang/Lex/ |
H A D | PreprocessingRecord.h | 43 unsigned alignment = 8) noexcept; 112 unsigned alignment = 8) noexcept { 113 return ::operator new(bytes, PR, alignment); 119 unsigned alignment) noexcept { 120 return ::operator delete(ptr, PR, alignment); 572 unsigned alignment) noexcept { 573 return PR.Allocate(bytes, alignment);
|
/freebsd-current/sys/arm64/arm64/ |
H A D | busdma_bounce.c | 123 #define dmat_alignment(dmat) ((dmat)->common.alignment) 198 * Return true if the given address does not fall on the alignment boundary. 204 return (!vm_addr_align_ok(addr, dmat->common.alignment)); 247 bounce_bus_dma_tag_create(bus_dma_tag_t parent, bus_size_t alignment, argument 257 NULL, alignment, boundary, lowaddr, highaddr, maxsize, nsegments, 280 newtag->alloc_alignment = newtag->common.alignment; 289 newtag->alloc_alignment = MAX(newtag->common.alignment, 296 newtag->common.alignment > 1) 410 * exclusion region, a data alignment that is stricter than 1, and/or 423 * limit. Even if the tag isn't subject of bouncing due to alignment [all...] |