Searched refs:bo (Results 1 - 25 of 87) sorted by relevance

1234

/freebsd-11-stable/sys/sys/
H A Dbufobj.h86 #define BO_STRATEGY(bo, bp) ((bo)->bo_ops->bop_strategy((bo), (bp)))
87 #define BO_SYNC(bo, w) ((bo)->bo_ops->bop_sync((bo), (w)))
88 #define BO_WRITE(bo, bp) ((bo)->bo_ops->bop_write((bp)))
89 #define BO_BDFLUSH(bo, bp) ((bo)
[all...]
/freebsd-11-stable/sys/dev/drm2/radeon/
H A Dradeon_object.c48 static void radeon_bo_clear_surface_reg(struct radeon_bo *bo);
55 static void radeon_bo_clear_va(struct radeon_bo *bo) argument
59 list_for_each_entry_safe(bo_va, tmp, &bo->va, bo_list) {
61 radeon_vm_bo_rmv(bo->rdev, bo_va);
67 struct radeon_bo *bo; local
69 bo = container_of(tbo, struct radeon_bo, tbo);
70 sx_xlock(&bo->rdev->gem.mutex);
71 list_del_init(&bo->list);
72 sx_xunlock(&bo->rdev->gem.mutex);
73 radeon_bo_clear_surface_reg(bo);
79 radeon_ttm_bo_is_radeon_bo(struct ttm_buffer_object *bo) argument
121 struct radeon_bo *bo; local
174 radeon_bo_kmap(struct radeon_bo *bo, void **ptr) argument
197 radeon_bo_kunmap(struct radeon_bo *bo) argument
206 radeon_bo_unref(struct radeon_bo **bo) argument
222 radeon_bo_pin_restricted(struct radeon_bo *bo, u32 domain, u64 max_offset, u64 *gpu_addr) argument
278 radeon_bo_pin(struct radeon_bo *bo, u32 domain, u64 *gpu_addr) argument
283 radeon_bo_unpin(struct radeon_bo *bo) argument
315 struct radeon_bo *bo, *n; local
366 struct radeon_bo *bo; local
398 radeon_bo_fbdev_mmap(struct radeon_bo *bo, struct vm_area_struct *vma) argument
405 radeon_bo_get_surface_reg(struct radeon_bo *bo) argument
461 radeon_bo_clear_surface_reg(struct radeon_bo *bo) argument
476 radeon_bo_set_tiling_flags(struct radeon_bo *bo, uint32_t tiling_flags, uint32_t pitch) argument
536 radeon_bo_get_tiling_flags(struct radeon_bo *bo, uint32_t *tiling_flags, uint32_t *pitch) argument
548 radeon_bo_check_tiling(struct radeon_bo *bo, bool has_moved, bool force_drop) argument
577 radeon_bo_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) argument
588 radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo) argument
619 radeon_bo_wait(struct radeon_bo *bo, u32 *mem_type, bool no_wait) argument
646 radeon_bo_reserve(struct radeon_bo *bo, bool no_intr) argument
[all...]
H A Dradeon_object.h65 int radeon_bo_reserve(struct radeon_bo *bo, bool no_intr);
67 static inline void radeon_bo_unreserve(struct radeon_bo *bo) argument
69 ttm_bo_unreserve(&bo->tbo);
73 * radeon_bo_gpu_offset - return GPU offset of bo
74 * @bo: radeon object for which we query the offset
81 static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) argument
83 return bo->tbo.offset;
86 static inline unsigned long radeon_bo_size(struct radeon_bo *bo) argument
88 return bo->tbo.num_pages << PAGE_SHIFT;
91 static inline bool radeon_bo_is_reserved(struct radeon_bo *bo) argument
96 radeon_bo_ngpu_pages(struct radeon_bo *bo) argument
101 radeon_bo_gpu_page_alignment(struct radeon_bo *bo) argument
115 radeon_bo_mmap_offset(struct radeon_bo *bo) argument
[all...]
H A Dradeon_prime.c40 struct radeon_bo *bo = attachment->dmabuf->priv; local
41 struct drm_device *dev = bo->rdev->ddev;
42 int npages = bo->tbo.num_pages;
47 sg = drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages);
63 struct radeon_bo *bo = dma_buf->priv; local
65 if (bo->gem_base.export_dma_buf == dma_buf) {
66 DRM_ERROR("unreference dmabuf %p\n", &bo->gem_base);
67 bo->gem_base.export_dma_buf = NULL;
68 drm_gem_object_unreference_unlocked(&bo->gem_base);
98 struct radeon_bo *bo local
122 struct radeon_bo *bo = dma_buf->priv; local
151 struct radeon_bo *bo; local
172 struct radeon_bo *bo = gem_to_radeon_bo(obj); local
194 struct radeon_bo *bo; local
[all...]
H A Dradeon_trace.h19 TP_PROTO(struct radeon_bo *bo),
20 TP_ARGS(bo),
22 __field(struct radeon_bo *, bo)
27 __entry->bo = bo;
28 __entry->pages = bo->tbo.num_pages;
30 TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages)
H A Dradeon_sa.c32 * We store the last allocated bo in "hole", we always try to allocate
33 * after the last allocated bo. Principle is that in a linear GPU ring
34 * progression was is after last is the oldest bo we allocated and thus
37 * If it's not the case we skip over the bo after last to the closest
38 * done bo if such one exist. If none exist and we are not asked to
62 sa_manager->bo = NULL;
72 RADEON_GEM_DOMAIN_CPU, NULL, &sa_manager->bo);
74 dev_err(rdev->dev, "(%d) failed to allocate bo for manager\n", r);
96 radeon_bo_unref(&sa_manager->bo);
107 if (sa_manager->bo
[all...]
H A Dradeon_ttm.c175 static void radeon_evict_flags(struct ttm_buffer_object *bo, argument
181 if (!radeon_ttm_bo_is_radeon_bo(bo)) {
190 rbo = container_of(bo, struct radeon_bo, tbo);
191 switch (bo->mem.mem_type) {
205 static int radeon_verify_access(struct ttm_buffer_object *bo) argument
210 static void radeon_move_null(struct ttm_buffer_object *bo, argument
213 struct ttm_mem_reg *old_mem = &bo->mem;
220 static int radeon_move_blit(struct ttm_buffer_object *bo, argument
230 rdev = radeon_get_rdev(bo->bdev);
265 fence = bo
276 radeon_move_vram_ram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument
323 radeon_move_ram_vram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument
363 radeon_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument
805 struct ttm_buffer_object *bo; local
[all...]
/freebsd-11-stable/sys/dev/drm2/ttm/
H A Dttm_bo.c44 static int ttm_bo_setup_vm(struct ttm_buffer_object *bo);
77 static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, argument
83 bo, bo->mem.num_pages, bo->mem.size >> 10,
84 bo->mem.size >> 20);
92 ttm_mem_type_debug(bo->bdev, mem_type);
111 static void ttm_bo_release_list(struct ttm_buffer_object *bo) argument
113 struct ttm_bo_device *bdev = bo->bdev;
114 size_t acc_size = bo
136 ttm_bo_wait_unreserved_locked(struct ttm_buffer_object *bo, bool interruptible) argument
159 ttm_bo_add_to_lru(struct ttm_buffer_object *bo) argument
181 ttm_bo_del_from_lru(struct ttm_buffer_object *bo) argument
202 ttm_bo_reserve_nolru(struct ttm_buffer_object *bo, bool interruptible, bool no_wait, bool use_sequence, uint32_t sequence) argument
266 ttm_bo_list_ref_sub(struct ttm_buffer_object *bo, int count, bool never_free) argument
279 ttm_bo_reserve(struct ttm_buffer_object *bo, bool interruptible, bool no_wait, bool use_sequence, uint32_t sequence) argument
300 ttm_bo_reserve_slowpath_nolru(struct ttm_buffer_object *bo, bool interruptible, uint32_t sequence) argument
334 ttm_bo_reserve_slowpath(struct ttm_buffer_object *bo, bool interruptible, uint32_t sequence) argument
351 ttm_bo_unreserve_locked(struct ttm_buffer_object *bo) argument
358 ttm_bo_unreserve(struct ttm_buffer_object *bo) argument
370 ttm_bo_add_ttm(struct ttm_buffer_object *bo, bool zero_alloc) argument
412 ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem, bool evict, bool interruptible, bool no_wait_gpu) argument
525 ttm_bo_cleanup_memtype_use(struct ttm_buffer_object *bo) argument
550 ttm_bo_cleanup_refs_or_queue(struct ttm_buffer_object *bo) argument
608 ttm_bo_cleanup_refs_and_unlock(struct ttm_buffer_object *bo, bool interruptible, bool no_wait_gpu) argument
758 ttm_bo_release(struct ttm_buffer_object *bo) argument
781 struct ttm_buffer_object *bo = *p_bo; local
805 ttm_bo_evict(struct ttm_buffer_object *bo, bool interruptible, bool no_wait_gpu) argument
867 struct ttm_buffer_object *bo; local
907 ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) argument
919 ttm_bo_mem_force_space(struct ttm_buffer_object *bo, uint32_t mem_type, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument
999 ttm_bo_mem_space(struct ttm_buffer_object *bo, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) argument
1106 ttm_bo_move_buffer(struct ttm_buffer_object *bo, struct ttm_placement *placement, bool interruptible, bool no_wait_gpu) argument
1167 ttm_bo_validate(struct ttm_buffer_object *bo, struct ttm_placement *placement, bool interruptible, bool no_wait_gpu) argument
1208 ttm_bo_check_placement(struct ttm_buffer_object *bo, struct ttm_placement *placement) argument
1217 ttm_bo_init(struct ttm_bo_device *bdev, struct ttm_buffer_object *bo, unsigned long size, enum ttm_bo_type type, struct ttm_placement *placement, uint32_t page_alignment, bool interruptible, struct vm_object *persistent_swap_storage, size_t acc_size, struct sg_table *sg, void (*destroy) (struct ttm_buffer_object *)) argument
1347 struct ttm_buffer_object *bo; local
1647 ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) argument
1654 ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) argument
1664 ttm_bo_vm_insert_rb(struct ttm_buffer_object *bo) argument
1683 ttm_bo_setup_vm(struct ttm_buffer_object *bo) argument
1720 ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, bool interruptible, bool no_wait) argument
1774 ttm_bo_synccpu_write_grab(struct ttm_buffer_object *bo, bool no_wait) argument
1795 ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo) argument
1809 struct ttm_buffer_object *bo; local
[all...]
H A Dttm_execbuf_util.c41 struct ttm_buffer_object *bo = entry->bo; local
46 ttm_bo_add_to_lru(bo);
51 atomic_set(&bo->reserved, 0);
52 wakeup(bo);
61 struct ttm_buffer_object *bo = entry->bo; local
66 entry->put_count = ttm_bo_del_from_lru(bo);
77 struct ttm_buffer_object *bo = entry->bo; local
137 struct ttm_buffer_object *bo = entry->bo; local
208 struct ttm_buffer_object *bo; local
[all...]
H A Dttm_bo_vm.c76 struct ttm_buffer_object *bo; local
79 bo = RB_ROOT(&bdev->addr_space_rb);
80 while (bo != NULL) {
81 cur_offset = bo->vm_node->start;
83 best_bo = bo;
86 bo = RB_RIGHT(bo, vm_rb);
88 bo = RB_LEFT(bo, vm_rb);
106 struct ttm_buffer_object *bo local
311 struct ttm_buffer_object *bo = handle; local
327 struct ttm_buffer_object *bo; local
369 ttm_bo_release_mmap(struct ttm_buffer_object *bo) argument
[all...]
H A Dttm_bo_util.c39 void ttm_bo_free_old_node(struct ttm_buffer_object *bo) argument
41 ttm_bo_mem_put(bo, &bo->mem);
44 int ttm_bo_move_ttm(struct ttm_buffer_object *bo, argument
48 struct ttm_tt *ttm = bo->ttm;
49 struct ttm_mem_reg *old_mem = &bo->mem;
54 ttm_bo_free_old_node(bo);
102 struct ttm_buffer_object *bo; local
107 bo = list_first_entry(&man->io_reserve_lru,
110 list_del_init(&bo
155 ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) argument
175 ttm_mem_io_free_vm(struct ttm_buffer_object *bo) argument
293 ttm_bo_move_memcpy(struct ttm_buffer_object *bo, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument
381 ttm_transfered_destroy(struct ttm_buffer_object *bo) argument
402 ttm_buffer_object_transfer(struct ttm_buffer_object *bo, struct ttm_buffer_object **new_obj) argument
457 ttm_bo_ioremap(struct ttm_buffer_object *bo, unsigned long offset, unsigned long size, struct ttm_bo_kmap_obj *map) argument
478 ttm_bo_kmap_ttm(struct ttm_buffer_object *bo, unsigned long start_page, unsigned long num_pages, struct ttm_bo_kmap_obj *map) argument
529 ttm_bo_kmap(struct ttm_buffer_object *bo, unsigned long start_page, unsigned long num_pages, struct ttm_bo_kmap_obj *map) argument
565 struct ttm_buffer_object *bo = map->bo; local
596 ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, void *sync_obj, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) argument
[all...]
H A Dttm_bo_api.h138 * @type: The bo type.
209 * Members protected by the bo::reserved lock.
218 * Members protected by the bo::reserved lock only when written to.
244 * than NULL requires bo::reserved to be held. This allows for
261 * and the bo::lock when written to. Can be read with
278 * Object describing a kernel mapping. Since a TTM bo may be located
297 struct ttm_buffer_object *bo; member in struct:ttm_bo_kmap_obj
303 * @bo: The buffer object.
309 ttm_bo_reference(struct ttm_buffer_object *bo) argument
311 refcount_acquire(&bo
735 ttm_bo_is_reserved(struct ttm_buffer_object *bo) argument
[all...]
H A Dttm_bo_driver.h181 * @bo: Pointer to the buffer object we're allocating space for.
205 struct ttm_buffer_object *bo,
374 * @bo: the buffer object to be evicted
376 * Return the bo flags for a buffer which is not mapped to the hardware.
378 * finished, they'll end up in bo->mem.flags
381 void(*evict_flags) (struct ttm_buffer_object *bo,
386 * @bo: the buffer to move
396 int (*move) (struct ttm_buffer_object *bo,
404 * @bo: Pointer to a buffer object.
414 int (*verify_access) (struct ttm_buffer_object *bo);
[all...]
H A Dttm_execbuf_util.h41 * @bo: refcounted buffer object pointer.
42 * @reserved: Indicates whether @bo has been reserved for validation.
43 * @removed: Indicates whether @bo has been removed from lru lists.
44 * @put_count: Number of outstanding references on bo::list_kref.
50 struct ttm_buffer_object *bo; member in struct:ttm_validate_buffer
/freebsd-11-stable/sys/arm/nvidia/drm2/
H A Dtegra_bo.c51 tegra_bo_destruct(struct tegra_bo *bo) argument
57 if (bo->cdev_pager == NULL)
60 size = round_page(bo->gem_obj.size);
61 if (bo->vbase != 0)
62 pmap_qremove(bo->vbase, bo->npages);
64 VM_OBJECT_WLOCK(bo->cdev_pager);
65 for (i = 0; i < bo->npages; i++) {
66 m = bo->m[i];
67 cdev_pager_free_page(bo
84 struct tegra_bo *bo; local
136 tegra_bo_init_pager(struct tegra_bo *bo) argument
175 tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo) argument
203 struct tegra_bo *bo; local
243 struct tegra_bo *bo; local
267 struct tegra_bo *bo; local
[all...]
H A Dtegra_fb.c49 struct tegra_bo *bo; local
54 bo = fb->planes[i];
55 if (bo != NULL)
56 drm_gem_object_unreference_unlocked(&bo->gem_obj);
125 struct tegra_bo *bo; local
152 rv = tegra_bo_create(drm_dev, size, &bo);
165 rv = fb_alloc(drm_dev, &mode_cmd, &bo, 1, &fb);
175 info->fb_vbase = bo->vbase;
176 info->fb_pbase = bo->pbase;
183 DRM_DEBUG_KMS("allocated %dx%d (s %dbits) fb size: %d, bo
[all...]
/freebsd-11-stable/sys/geom/
H A Dgeom_vfs.h38 void g_vfs_strategy(struct bufobj *bo, struct buf *bp);
H A Dgeom_vfs.c161 g_vfs_strategy(struct bufobj *bo, struct buf *bp) argument
167 cp = bo->bo_private;
231 struct bufobj *bo; local
237 bo = &vp->v_bufobj;
238 if (bo->bo_private != vp)
247 sc->sc_bo = bo;
260 bo->bo_ops = g_vfs_bufops;
261 bo->bo_private = cp;
262 bo->bo_bsize = pp->sectorsize;
/freebsd-11-stable/sys/fs/msdosfs/
H A Dmsdosfs_fat.c134 u_long bo; local
197 fatblock(pmp, byteoffset, &bn, &bsize, &bo);
209 if (bo >= bsize) {
215 cn = getulong(&bp->b_data[bo]);
217 cn = getushort(&bp->b_data[bo]);
466 u_long bn, bo, bsize, byteoffset; local
504 fatblock(pmp, byteoffset, &bn, &bsize, &bo);
513 readcn = getulong(&bp->b_data[bo]);
515 readcn = getushort(&bp->b_data[bo]);
527 readcn = getushort(&bp->b_data[bo]);
572 u_long bn, bo, bsize, byteoffset, readcn, newc; local
834 u_long bn, bo, bsize, byteoffset; local
895 u_long bn, bo, bsize, byteoffset, cn, readcn; local
1134 u_long bn, bo, bsize, byteoffset, fatval; local
[all...]
/freebsd-11-stable/sys/kern/
H A Dvfs_subr.c103 static int flushbuflist(struct bufv *bufv, int flags, struct bufobj *bo,
265 * bo->bo_synclist
464 struct bufobj *bo; local
481 bo = &vp->v_bufobj;
482 bo->__bo_vnode = vp;
483 rw_init(BO_LOCKPTR(bo), "bufobj interlock");
484 bo->bo_private = vp;
485 TAILQ_INIT(&bo->bo_clean.bv_hd);
486 TAILQ_INIT(&bo->bo_dirty.bv_hd);
506 struct bufobj *bo; local
1669 bufobj_invalbuf(struct bufobj *bo, int flags, int slpflag, int slptimeo) argument
1774 flushbuflist(struct bufv *bufv, int flags, struct bufobj *bo, int slpflag, int slptimeo) argument
1835 bnoreuselist(struct bufv *bufv, struct bufobj *bo, daddr_t startn, daddr_t endn) argument
1889 struct bufobj *bo; local
2011 buf_vlist_add(struct buf *bp, struct bufobj *bo, b_xflags_t xflags) argument
2050 gbincore(struct bufobj *bo, daddr_t lblkno) argument
2067 struct bufobj *bo; local
2092 struct bufobj *bo; local
2125 vn_syncer_add_to_worklist(struct bufobj *bo, int delay) argument
2172 sync_vnode(struct synclist *slp, struct bufobj **bo, struct thread *td) argument
2226 struct bufobj *bo; local
2427 struct bufobj *bo; local
2905 struct bufobj *bo; local
4192 struct bufobj *bo; local
4266 struct bufobj *bo; local
4321 struct bufobj *bo; local
[all...]
H A Dvfs_default.c647 struct bufobj *bo; local
660 bo = &vp->v_bufobj;
661 BO_LOCK(bo);
666 TAILQ_FOREACH(bp, &bo->bo_dirty.bv_hd, b_bobufs) {
675 TAILQ_FOREACH_SAFE(bp, &bo->bo_dirty.bv_hd, b_bobufs, nbp) {
684 BO_LOCKPTR(bo)) != 0) {
685 BO_LOCK(bo);
688 BO_LOCK(bo);
690 BO_UNLOCK(bo);
691 KASSERT(bp->b_bufobj == bo,
1075 struct bufobj *bo; local
[all...]
/freebsd-11-stable/sys/geom/label/
H A Dg_label_gpt.c58 u_int bo; local
62 bo = LITTLE_ENDIAN; /* GPT is little-endian */
64 ch = (bo == BIG_ENDIAN) ? be16toh(*str) : le16toh(*str);
68 c = (bo == BIG_ENDIAN) ? be16toh(*str)
79 bo = (bo == BIG_ENDIAN) ? LITTLE_ENDIAN : BIG_ENDIAN;
/freebsd-11-stable/sys/ufs/ffs/
H A Dffs_rawread.c100 struct bufobj *bo; local
105 bo = &vp->v_bufobj;
106 BO_LOCK(bo);
108 if (bo->bo_numoutput > 0 ||
109 bo->bo_dirty.bv_cnt > 0 ||
113 BO_UNLOCK(bo);
151 BO_LOCK(bo);
155 BO_UNLOCK(bo);
162 if (bo->bo_dirty.bv_cnt > 0) {
163 BO_UNLOCK(bo);
[all...]
/freebsd-11-stable/crypto/openssl/crypto/perlasm/
H A Dppc-xlate.pl118 my $bo = $f=~/[\+\-]/ ? 16+9 : 16; # optional "to be taken" hint
119 " bc $bo,0,".shift;
123 my $bo = $f=~/\-/ ? 12+2 : 12; # optional "not to be taken" hint
125 " .long ".sprintf "0x%x",19<<26|$bo<<21|16<<1 :
126 " bclr $bo,0";
130 my $bo = $f=~/\-/ ? 4+2 : 4; # optional "not to be taken" hint
132 " .long ".sprintf "0x%x",19<<26|$bo<<21|2<<16|16<<1 :
133 " bclr $bo,2";
137 my $bo = $f=~/-/ ? 12+2 : 12; # optional "not to be taken" hint
139 " .long ".sprintf "0x%X",19<<26|$bo<<2
[all...]
/freebsd-11-stable/sys/net80211/
H A Dieee80211_output.c2902 struct ieee80211_beacon_offsets *bo = &vap->iv_bcn_off; local
2937 memset(bo, 0, sizeof(*bo));
2944 bo->bo_caps = (uint16_t *)frm;
2961 bo->bo_cfp = frm;
2964 bo->bo_tim = frm;
2969 bo->bo_tim_len = 0;
2982 bo->bo_tim_len = 1;
2984 bo->bo_tim_trailer = frm;
2991 bo
3167 struct ieee80211_beacon_offsets *bo = &vap->iv_bcn_off; local
[all...]

Completed in 219 milliseconds

1234