/freebsd-11-stable/sys/sys/ |
H A D | bufobj.h | 86 #define BO_STRATEGY(bo, bp) ((bo)->bo_ops->bop_strategy((bo), (bp))) 87 #define BO_SYNC(bo, w) ((bo)->bo_ops->bop_sync((bo), (w))) 88 #define BO_WRITE(bo, bp) ((bo)->bo_ops->bop_write((bp))) 89 #define BO_BDFLUSH(bo, bp) ((bo) [all...] |
/freebsd-11-stable/sys/dev/drm2/radeon/ |
H A D | radeon_object.c | 48 static void radeon_bo_clear_surface_reg(struct radeon_bo *bo); 55 static void radeon_bo_clear_va(struct radeon_bo *bo) argument 59 list_for_each_entry_safe(bo_va, tmp, &bo->va, bo_list) { 61 radeon_vm_bo_rmv(bo->rdev, bo_va); 67 struct radeon_bo *bo; local 69 bo = container_of(tbo, struct radeon_bo, tbo); 70 sx_xlock(&bo->rdev->gem.mutex); 71 list_del_init(&bo->list); 72 sx_xunlock(&bo->rdev->gem.mutex); 73 radeon_bo_clear_surface_reg(bo); 79 radeon_ttm_bo_is_radeon_bo(struct ttm_buffer_object *bo) argument 121 struct radeon_bo *bo; local 174 radeon_bo_kmap(struct radeon_bo *bo, void **ptr) argument 197 radeon_bo_kunmap(struct radeon_bo *bo) argument 206 radeon_bo_unref(struct radeon_bo **bo) argument 222 radeon_bo_pin_restricted(struct radeon_bo *bo, u32 domain, u64 max_offset, u64 *gpu_addr) argument 278 radeon_bo_pin(struct radeon_bo *bo, u32 domain, u64 *gpu_addr) argument 283 radeon_bo_unpin(struct radeon_bo *bo) argument 315 struct radeon_bo *bo, *n; local 366 struct radeon_bo *bo; local 398 radeon_bo_fbdev_mmap(struct radeon_bo *bo, struct vm_area_struct *vma) argument 405 radeon_bo_get_surface_reg(struct radeon_bo *bo) argument 461 radeon_bo_clear_surface_reg(struct radeon_bo *bo) argument 476 radeon_bo_set_tiling_flags(struct radeon_bo *bo, uint32_t tiling_flags, uint32_t pitch) argument 536 radeon_bo_get_tiling_flags(struct radeon_bo *bo, uint32_t *tiling_flags, uint32_t *pitch) argument 548 radeon_bo_check_tiling(struct radeon_bo *bo, bool has_moved, bool force_drop) argument 577 radeon_bo_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) argument 588 radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo) argument 619 radeon_bo_wait(struct radeon_bo *bo, u32 *mem_type, bool no_wait) argument 646 radeon_bo_reserve(struct radeon_bo *bo, bool no_intr) argument [all...] |
H A D | radeon_object.h | 65 int radeon_bo_reserve(struct radeon_bo *bo, bool no_intr); 67 static inline void radeon_bo_unreserve(struct radeon_bo *bo) argument 69 ttm_bo_unreserve(&bo->tbo); 73 * radeon_bo_gpu_offset - return GPU offset of bo 74 * @bo: radeon object for which we query the offset 81 static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) argument 83 return bo->tbo.offset; 86 static inline unsigned long radeon_bo_size(struct radeon_bo *bo) argument 88 return bo->tbo.num_pages << PAGE_SHIFT; 91 static inline bool radeon_bo_is_reserved(struct radeon_bo *bo) argument 96 radeon_bo_ngpu_pages(struct radeon_bo *bo) argument 101 radeon_bo_gpu_page_alignment(struct radeon_bo *bo) argument 115 radeon_bo_mmap_offset(struct radeon_bo *bo) argument [all...] |
H A D | radeon_prime.c | 40 struct radeon_bo *bo = attachment->dmabuf->priv; local 41 struct drm_device *dev = bo->rdev->ddev; 42 int npages = bo->tbo.num_pages; 47 sg = drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages); 63 struct radeon_bo *bo = dma_buf->priv; local 65 if (bo->gem_base.export_dma_buf == dma_buf) { 66 DRM_ERROR("unreference dmabuf %p\n", &bo->gem_base); 67 bo->gem_base.export_dma_buf = NULL; 68 drm_gem_object_unreference_unlocked(&bo->gem_base); 98 struct radeon_bo *bo local 122 struct radeon_bo *bo = dma_buf->priv; local 151 struct radeon_bo *bo; local 172 struct radeon_bo *bo = gem_to_radeon_bo(obj); local 194 struct radeon_bo *bo; local [all...] |
H A D | radeon_trace.h | 19 TP_PROTO(struct radeon_bo *bo), 20 TP_ARGS(bo), 22 __field(struct radeon_bo *, bo) 27 __entry->bo = bo; 28 __entry->pages = bo->tbo.num_pages; 30 TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages)
|
H A D | radeon_sa.c | 32 * We store the last allocated bo in "hole", we always try to allocate 33 * after the last allocated bo. Principle is that in a linear GPU ring 34 * progression was is after last is the oldest bo we allocated and thus 37 * If it's not the case we skip over the bo after last to the closest 38 * done bo if such one exist. If none exist and we are not asked to 62 sa_manager->bo = NULL; 72 RADEON_GEM_DOMAIN_CPU, NULL, &sa_manager->bo); 74 dev_err(rdev->dev, "(%d) failed to allocate bo for manager\n", r); 96 radeon_bo_unref(&sa_manager->bo); 107 if (sa_manager->bo [all...] |
H A D | radeon_ttm.c | 175 static void radeon_evict_flags(struct ttm_buffer_object *bo, argument 181 if (!radeon_ttm_bo_is_radeon_bo(bo)) { 190 rbo = container_of(bo, struct radeon_bo, tbo); 191 switch (bo->mem.mem_type) { 205 static int radeon_verify_access(struct ttm_buffer_object *bo) argument 210 static void radeon_move_null(struct ttm_buffer_object *bo, argument 213 struct ttm_mem_reg *old_mem = &bo->mem; 220 static int radeon_move_blit(struct ttm_buffer_object *bo, argument 230 rdev = radeon_get_rdev(bo->bdev); 265 fence = bo 276 radeon_move_vram_ram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument 323 radeon_move_ram_vram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument 363 radeon_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument 805 struct ttm_buffer_object *bo; local [all...] |
/freebsd-11-stable/sys/dev/drm2/ttm/ |
H A D | ttm_bo.c | 44 static int ttm_bo_setup_vm(struct ttm_buffer_object *bo); 77 static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, argument 83 bo, bo->mem.num_pages, bo->mem.size >> 10, 84 bo->mem.size >> 20); 92 ttm_mem_type_debug(bo->bdev, mem_type); 111 static void ttm_bo_release_list(struct ttm_buffer_object *bo) argument 113 struct ttm_bo_device *bdev = bo->bdev; 114 size_t acc_size = bo 136 ttm_bo_wait_unreserved_locked(struct ttm_buffer_object *bo, bool interruptible) argument 159 ttm_bo_add_to_lru(struct ttm_buffer_object *bo) argument 181 ttm_bo_del_from_lru(struct ttm_buffer_object *bo) argument 202 ttm_bo_reserve_nolru(struct ttm_buffer_object *bo, bool interruptible, bool no_wait, bool use_sequence, uint32_t sequence) argument 266 ttm_bo_list_ref_sub(struct ttm_buffer_object *bo, int count, bool never_free) argument 279 ttm_bo_reserve(struct ttm_buffer_object *bo, bool interruptible, bool no_wait, bool use_sequence, uint32_t sequence) argument 300 ttm_bo_reserve_slowpath_nolru(struct ttm_buffer_object *bo, bool interruptible, uint32_t sequence) argument 334 ttm_bo_reserve_slowpath(struct ttm_buffer_object *bo, bool interruptible, uint32_t sequence) argument 351 ttm_bo_unreserve_locked(struct ttm_buffer_object *bo) argument 358 ttm_bo_unreserve(struct ttm_buffer_object *bo) argument 370 ttm_bo_add_ttm(struct ttm_buffer_object *bo, bool zero_alloc) argument 412 ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem, bool evict, bool interruptible, bool no_wait_gpu) argument 525 ttm_bo_cleanup_memtype_use(struct ttm_buffer_object *bo) argument 550 ttm_bo_cleanup_refs_or_queue(struct ttm_buffer_object *bo) argument 608 ttm_bo_cleanup_refs_and_unlock(struct ttm_buffer_object *bo, bool interruptible, bool no_wait_gpu) argument 758 ttm_bo_release(struct ttm_buffer_object *bo) argument 781 struct ttm_buffer_object *bo = *p_bo; local 805 ttm_bo_evict(struct ttm_buffer_object *bo, bool interruptible, bool no_wait_gpu) argument 867 struct ttm_buffer_object *bo; local 907 ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) argument 919 ttm_bo_mem_force_space(struct ttm_buffer_object *bo, uint32_t mem_type, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument 999 ttm_bo_mem_space(struct ttm_buffer_object *bo, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument 1106 ttm_bo_move_buffer(struct ttm_buffer_object *bo, struct ttm_placement *placement, bool interruptible, bool no_wait_gpu) argument 1167 ttm_bo_validate(struct ttm_buffer_object *bo, struct ttm_placement *placement, bool interruptible, bool no_wait_gpu) argument 1208 ttm_bo_check_placement(struct ttm_buffer_object *bo, struct ttm_placement *placement) argument 1217 ttm_bo_init(struct ttm_bo_device *bdev, struct ttm_buffer_object *bo, unsigned long size, enum ttm_bo_type type, struct ttm_placement *placement, uint32_t page_alignment, bool interruptible, struct vm_object *persistent_swap_storage, size_t acc_size, struct sg_table *sg, void (*destroy) (struct ttm_buffer_object *)) argument 1347 struct ttm_buffer_object *bo; local 1647 ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) argument 1654 ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) argument 1664 ttm_bo_vm_insert_rb(struct ttm_buffer_object *bo) argument 1683 ttm_bo_setup_vm(struct ttm_buffer_object *bo) argument 1720 ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, bool interruptible, bool no_wait) argument 1774 ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait) argument 1795 ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo) argument 1809 struct ttm_buffer_object *bo; local [all...] |
H A D | ttm_execbuf_util.c | 41 struct ttm_buffer_object *bo = entry->bo; local 46 ttm_bo_add_to_lru(bo); 51 atomic_set(&bo->reserved, 0); 52 wakeup(bo); 61 struct ttm_buffer_object *bo = entry->bo; local 66 entry->put_count = ttm_bo_del_from_lru(bo); 77 struct ttm_buffer_object *bo = entry->bo; local 137 struct ttm_buffer_object *bo = entry->bo; local 208 struct ttm_buffer_object *bo; local [all...] |
H A D | ttm_bo_vm.c | 76 struct ttm_buffer_object *bo; local 79 bo = RB_ROOT(&bdev->addr_space_rb); 80 while (bo != NULL) { 81 cur_offset = bo->vm_node->start; 83 best_bo = bo; 86 bo = RB_RIGHT(bo, vm_rb); 88 bo = RB_LEFT(bo, vm_rb); 106 struct ttm_buffer_object *bo local 311 struct ttm_buffer_object *bo = handle; local 327 struct ttm_buffer_object *bo; local 369 ttm_bo_release_mmap(struct ttm_buffer_object *bo) argument [all...] |
H A D | ttm_bo_util.c | 39 void ttm_bo_free_old_node(struct ttm_buffer_object *bo) argument 41 ttm_bo_mem_put(bo, &bo->mem); 44 int ttm_bo_move_ttm(struct ttm_buffer_object *bo, argument 48 struct ttm_tt *ttm = bo->ttm; 49 struct ttm_mem_reg *old_mem = &bo->mem; 54 ttm_bo_free_old_node(bo); 102 struct ttm_buffer_object *bo; local 107 bo = list_first_entry(&man->io_reserve_lru, 110 list_del_init(&bo 155 ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) argument 175 ttm_mem_io_free_vm(struct ttm_buffer_object *bo) argument 293 ttm_bo_move_memcpy(struct ttm_buffer_object *bo, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument 381 ttm_transfered_destroy(struct ttm_buffer_object *bo) argument 402 ttm_buffer_object_transfer(struct ttm_buffer_object *bo, struct ttm_buffer_object **new_obj) argument 457 ttm_bo_ioremap(struct ttm_buffer_object *bo, unsigned long offset, unsigned long size, struct ttm_bo_kmap_obj *map) argument 478 ttm_bo_kmap_ttm(struct ttm_buffer_object *bo, unsigned long start_page, unsigned long num_pages, struct ttm_bo_kmap_obj *map) argument 529 ttm_bo_kmap(struct ttm_buffer_object *bo, unsigned long start_page, unsigned long num_pages, struct ttm_bo_kmap_obj *map) argument 565 struct ttm_buffer_object *bo = map->bo; local 596 ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, void *sync_obj, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument [all...] |
H A D | ttm_bo_api.h | 138 * @type: The bo type. 209 * Members protected by the bo::reserved lock. 218 * Members protected by the bo::reserved lock only when written to. 244 * than NULL requires bo::reserved to be held. This allows for 261 * and the bo::lock when written to. Can be read with 278 * Object describing a kernel mapping. Since a TTM bo may be located 297 struct ttm_buffer_object *bo; member in struct:ttm_bo_kmap_obj 303 * @bo: The buffer object. 309 ttm_bo_reference(struct ttm_buffer_object *bo) argument 311 refcount_acquire(&bo 735 ttm_bo_is_reserved(struct ttm_buffer_object *bo) argument [all...] |
H A D | ttm_bo_driver.h | 181 * @bo: Pointer to the buffer object we're allocating space for. 205 struct ttm_buffer_object *bo, 374 * @bo: the buffer object to be evicted 376 * Return the bo flags for a buffer which is not mapped to the hardware. 378 * finished, they'll end up in bo->mem.flags 381 void(*evict_flags) (struct ttm_buffer_object *bo, 386 * @bo: the buffer to move 396 int (*move) (struct ttm_buffer_object *bo, 404 * @bo: Pointer to a buffer object. 414 int (*verify_access) (struct ttm_buffer_object *bo); [all...] |
H A D | ttm_execbuf_util.h | 41 * @bo: refcounted buffer object pointer. 42 * @reserved: Indicates whether @bo has been reserved for validation. 43 * @removed: Indicates whether @bo has been removed from lru lists. 44 * @put_count: Number of outstanding references on bo::list_kref. 50 struct ttm_buffer_object *bo; member in struct:ttm_validate_buffer
|
/freebsd-11-stable/sys/arm/nvidia/drm2/ |
H A D | tegra_bo.c | 51 tegra_bo_destruct(struct tegra_bo *bo) argument 57 if (bo->cdev_pager == NULL) 60 size = round_page(bo->gem_obj.size); 61 if (bo->vbase != 0) 62 pmap_qremove(bo->vbase, bo->npages); 64 VM_OBJECT_WLOCK(bo->cdev_pager); 65 for (i = 0; i < bo->npages; i++) { 66 m = bo->m[i]; 67 cdev_pager_free_page(bo 84 struct tegra_bo *bo; local 136 tegra_bo_init_pager(struct tegra_bo *bo) argument 175 tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo) argument 203 struct tegra_bo *bo; local 243 struct tegra_bo *bo; local 267 struct tegra_bo *bo; local [all...] |
H A D | tegra_fb.c | 49 struct tegra_bo *bo; local 54 bo = fb->planes[i]; 55 if (bo != NULL) 56 drm_gem_object_unreference_unlocked(&bo->gem_obj); 125 struct tegra_bo *bo; local 152 rv = tegra_bo_create(drm_dev, size, &bo); 165 rv = fb_alloc(drm_dev, &mode_cmd, &bo, 1, &fb); 175 info->fb_vbase = bo->vbase; 176 info->fb_pbase = bo->pbase; 183 DRM_DEBUG_KMS("allocated %dx%d (s %dbits) fb size: %d, bo [all...] |
/freebsd-11-stable/sys/geom/ |
H A D | geom_vfs.h | 38 void g_vfs_strategy(struct bufobj *bo, struct buf *bp);
|
H A D | geom_vfs.c | 161 g_vfs_strategy(struct bufobj *bo, struct buf *bp) argument 167 cp = bo->bo_private; 231 struct bufobj *bo; local 237 bo = &vp->v_bufobj; 238 if (bo->bo_private != vp) 247 sc->sc_bo = bo; 260 bo->bo_ops = g_vfs_bufops; 261 bo->bo_private = cp; 262 bo->bo_bsize = pp->sectorsize;
|
/freebsd-11-stable/sys/fs/msdosfs/ |
H A D | msdosfs_fat.c | 134 u_long bo; local 197 fatblock(pmp, byteoffset, &bn, &bsize, &bo); 209 if (bo >= bsize) { 215 cn = getulong(&bp->b_data[bo]); 217 cn = getushort(&bp->b_data[bo]); 466 u_long bn, bo, bsize, byteoffset; local 504 fatblock(pmp, byteoffset, &bn, &bsize, &bo); 513 readcn = getulong(&bp->b_data[bo]); 515 readcn = getushort(&bp->b_data[bo]); 527 readcn = getushort(&bp->b_data[bo]); 572 u_long bn, bo, bsize, byteoffset, readcn, newc; local 834 u_long bn, bo, bsize, byteoffset; local 895 u_long bn, bo, bsize, byteoffset, cn, readcn; local 1134 u_long bn, bo, bsize, byteoffset, fatval; local [all...] |
/freebsd-11-stable/sys/kern/ |
H A D | vfs_subr.c | 103 static int flushbuflist(struct bufv *bufv, int flags, struct bufobj *bo, 265 * bo->bo_synclist 464 struct bufobj *bo; local 481 bo = &vp->v_bufobj; 482 bo->__bo_vnode = vp; 483 rw_init(BO_LOCKPTR(bo), "bufobj interlock"); 484 bo->bo_private = vp; 485 TAILQ_INIT(&bo->bo_clean.bv_hd); 486 TAILQ_INIT(&bo->bo_dirty.bv_hd); 506 struct bufobj *bo; local 1669 bufobj_invalbuf(struct bufobj *bo, int flags, int slpflag, int slptimeo) argument 1774 flushbuflist(struct bufv *bufv, int flags, struct bufobj *bo, int slpflag, int slptimeo) argument 1835 bnoreuselist(struct bufv *bufv, struct bufobj *bo, daddr_t startn, daddr_t endn) argument 1889 struct bufobj *bo; local 2011 buf_vlist_add(struct buf *bp, struct bufobj *bo, b_xflags_t xflags) argument 2050 gbincore(struct bufobj *bo, daddr_t lblkno) argument 2067 struct bufobj *bo; local 2092 struct bufobj *bo; local 2125 vn_syncer_add_to_worklist(struct bufobj *bo, int delay) argument 2172 sync_vnode(struct synclist *slp, struct bufobj **bo, struct thread *td) argument 2226 struct bufobj *bo; local 2427 struct bufobj *bo; local 2905 struct bufobj *bo; local 4192 struct bufobj *bo; local 4266 struct bufobj *bo; local 4321 struct bufobj *bo; local [all...] |
H A D | vfs_default.c | 647 struct bufobj *bo; local 660 bo = &vp->v_bufobj; 661 BO_LOCK(bo); 666 TAILQ_FOREACH(bp, &bo->bo_dirty.bv_hd, b_bobufs) { 675 TAILQ_FOREACH_SAFE(bp, &bo->bo_dirty.bv_hd, b_bobufs, nbp) { 684 BO_LOCKPTR(bo)) != 0) { 685 BO_LOCK(bo); 688 BO_LOCK(bo); 690 BO_UNLOCK(bo); 691 KASSERT(bp->b_bufobj == bo, 1075 struct bufobj *bo; local [all...] |
/freebsd-11-stable/sys/geom/label/ |
H A D | g_label_gpt.c | 58 u_int bo; local 62 bo = LITTLE_ENDIAN; /* GPT is little-endian */ 64 ch = (bo == BIG_ENDIAN) ? be16toh(*str) : le16toh(*str); 68 c = (bo == BIG_ENDIAN) ? be16toh(*str) 79 bo = (bo == BIG_ENDIAN) ? LITTLE_ENDIAN : BIG_ENDIAN;
|
/freebsd-11-stable/sys/ufs/ffs/ |
H A D | ffs_rawread.c | 100 struct bufobj *bo; local 105 bo = &vp->v_bufobj; 106 BO_LOCK(bo); 108 if (bo->bo_numoutput > 0 || 109 bo->bo_dirty.bv_cnt > 0 || 113 BO_UNLOCK(bo); 151 BO_LOCK(bo); 155 BO_UNLOCK(bo); 162 if (bo->bo_dirty.bv_cnt > 0) { 163 BO_UNLOCK(bo); [all...] |
/freebsd-11-stable/crypto/openssl/crypto/perlasm/ |
H A D | ppc-xlate.pl | 118 my $bo = $f=~/[\+\-]/ ? 16+9 : 16; # optional "to be taken" hint 119 " bc $bo,0,".shift; 123 my $bo = $f=~/\-/ ? 12+2 : 12; # optional "not to be taken" hint 125 " .long ".sprintf "0x%x",19<<26|$bo<<21|16<<1 : 126 " bclr $bo,0"; 130 my $bo = $f=~/\-/ ? 4+2 : 4; # optional "not to be taken" hint 132 " .long ".sprintf "0x%x",19<<26|$bo<<21|2<<16|16<<1 : 133 " bclr $bo,2"; 137 my $bo = $f=~/-/ ? 12+2 : 12; # optional "not to be taken" hint 139 " .long ".sprintf "0x%X",19<<26|$bo<<2 [all...] |
/freebsd-11-stable/sys/net80211/ |
H A D | ieee80211_output.c | 2902 struct ieee80211_beacon_offsets *bo = &vap->iv_bcn_off; local 2937 memset(bo, 0, sizeof(*bo)); 2944 bo->bo_caps = (uint16_t *)frm; 2961 bo->bo_cfp = frm; 2964 bo->bo_tim = frm; 2969 bo->bo_tim_len = 0; 2982 bo->bo_tim_len = 1; 2984 bo->bo_tim_trailer = frm; 2991 bo 3167 struct ieee80211_beacon_offsets *bo = &vap->iv_bcn_off; local [all...] |