/linux-master/drivers/net/ethernet/brocade/bna/ |
H A D | bnad_debugfs.c | 310 void __iomem *rb, *reg_addr; local 335 rb = bfa_ioc_bar0(ioc); 348 reg_addr = rb + addr;
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.h | 67 struct rb_node rb; member in struct:amdgpu_bo_va_mapping
|
H A D | amdgpu_vm.c | 92 INTERVAL_TREE_DEFINE(struct amdgpu_bo_va_mapping, rb, uint64_t, __subtree_last, 2593 &vm->va.rb_root, rb) {
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 2679 rb_node = rb_prev(&node->rb); 2687 node = container_of(rb_node, struct interval_tree_node, rb); 2731 struct amdgpu_bo_va_mapping, rb); 3244 struct amdgpu_bo_va_mapping, rb);
|
/linux-master/fs/btrfs/ |
H A D | extent_map.c | 229 struct rb_node *rb; local 246 rb = rb_prev(&em->rb_node); 247 if (rb) 248 merge = rb_entry(rb, struct extent_map, rb_node); 249 if (rb && can_merge_extent_map(merge) && mergeable_maps(merge, em)) { 266 rb = rb_next(&em->rb_node); 267 if (rb) 268 merge = rb_entry(rb, struct extent_map, rb_node); 269 if (rb && can_merge_extent_map(merge) && mergeable_maps(em, merge)) {
|
/linux-master/fs/bcachefs/ |
H A D | btree_io.c | 1279 struct btree_read_bio *rb = local 1281 struct bch_fs *c = rb->c; 1282 struct btree *b = rb->b; 1283 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); 1284 struct bio *bio = &rb->bio; 1295 ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); 1296 rb->have_ioref = bch2_dev_get_ioref(ca, READ); 1298 bio->bi_iter.bi_sector = rb->pick.ptr.offset; 1301 if (rb->have_ioref) { 1313 if (rb 1361 struct btree_read_bio *rb = local 1557 struct btree_read_bio *rb = local 1606 struct btree_read_bio *rb = local 1649 struct btree_read_bio *rb; local [all...] |
/linux-master/drivers/firmware/arm_scmi/ |
H A D | raw_mode.c | 261 struct scmi_raw_buffer *rb = NULL; local 266 rb = list_first_entry(head, struct scmi_raw_buffer, node); 267 list_del_init(&rb->node); 271 return rb; 275 struct scmi_raw_buffer *rb) 280 rb->msg.len = rb->max_len; 283 list_add_tail(&rb->node, &q->free_bufs); 288 struct scmi_raw_buffer *rb) 293 list_add_tail(&rb 274 scmi_raw_buffer_put(struct scmi_raw_queue *q, struct scmi_raw_buffer *rb) argument 287 scmi_raw_buffer_enqueue(struct scmi_raw_queue *q, struct scmi_raw_buffer *rb) argument 302 struct scmi_raw_buffer *rb = NULL; local 315 struct scmi_raw_buffer *rb; local 326 struct scmi_raw_buffer *rb; local 698 struct scmi_raw_buffer *rb; local 741 struct scmi_raw_buffer *rb; local 1036 struct scmi_raw_buffer *rb; local 1317 struct scmi_raw_buffer *rb; local [all...] |
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a6xx_gpu.c | 1454 struct msm_ringbuffer *ring = gpu->rb[0]; 1483 struct msm_ringbuffer *ring = gpu->rb[0]; 1965 gpu_write64(gpu, REG_A6XX_CP_RB_BASE, gpu->rb[0]->iova); 1980 shadowptr(a6xx_gpu, gpu->rb[0])); 1986 rbmemptr(gpu->rb[0], bv_fence)); 1989 /* Always come up on rb 0 */ 1990 a6xx_gpu->cur_ring = gpu->rb[0]; 2010 OUT_PKT7(gpu->rb[0], CP_SET_SECURE_MODE, 1); 2011 OUT_RING(gpu->rb[0], 0x00000000); 2013 a6xx_flush(gpu, gpu->rb[ [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 404 static inline struct i915_priolist *to_priolist(struct rb_node *rb) argument 406 return rb_entry(rb, struct i915_priolist, node); 932 struct rb_node *rb; local 953 while ((rb = rb_first_cached(&sched_engine->queue))) { 954 struct i915_priolist *p = to_priolist(rb); 1924 struct rb_node *rb; local 1948 while ((rb = rb_first_cached(&sched_engine->queue))) { 1949 struct i915_priolist *p = to_priolist(rb); 5411 struct rb_node *rb; local 5427 for (rb [all...] |
/linux-master/sound/soc/sof/intel/ |
H A D | hda-stream.c | 867 PAGE_SIZE, &bus->rb); 964 if (bus->rb.area) 965 snd_dma_free_pages(&bus->rb);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_prime.c | 100 struct rb_node **p, *rb; local 110 rb = NULL; 115 rb = *p; 116 pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); 118 p = &rb->rb_right; 120 p = &rb->rb_left; 122 rb_link_node(&member->dmabuf_rb, rb, p); 125 rb = NULL; 130 rb = *p; 131 pos = rb_entry(rb, struc 146 struct rb_node *rb; local 168 struct rb_node *rb; local 191 struct rb_node *rb; local [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma.c | 152 struct rb_node *rb, **p; local 235 rb = NULL; 240 rb = *p; 241 pos = rb_entry(rb, struct i915_vma, obj_node); 250 p = &rb->rb_right; 252 p = &rb->rb_left; 256 rb_link_node(&vma->obj_node, rb, p); 289 struct rb_node *rb; local 291 rb = obj->vma.tree.rb_node; 292 while (rb) { [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_execlists_submission.c | 192 struct rb_node rb; member in struct:virtual_engine::ve_node 251 static struct i915_priolist *to_priolist(struct rb_node *rb) argument 253 return rb_entry(rb, struct i915_priolist, node); 281 struct rb_node *rb; local 283 rb = rb_first_cached(&sched_engine->queue); 284 if (!rb) 287 return to_priolist(rb)->priority; 292 struct rb_node *rb = rb_first_cached(&el->virtual); local 294 return rb ? rb_entry(rb, struc 1028 struct rb_node *rb = rb_first_cached(&el->virtual); local 1281 struct rb_node *rb; local 3152 struct rb_node *rb; local 3832 struct rb_node **parent, *rb; local 4093 struct rb_node *rb; local [all...] |
/linux-master/tools/testing/selftests/net/ |
H A D | tls.c | 1398 char rb[8001]; local 1403 res = recv(self->cfd, rb, 1404 left > sizeof(rb) ? sizeof(rb) : left, 0);
|
/linux-master/include/linux/ |
H A D | skbuff.h | 3961 #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode)
|
H A D | kernfs.h | 155 /* children rbtree starts here and goes through kn->rb */ 205 struct rb_node rb; member in struct:kernfs_node
|
H A D | fb.h | 766 extern int fb_find_mode_cvt(struct fb_videomode *mode, int margins, int rb);
|
/linux-master/kernel/printk/ |
H A D | printk.c | 1100 static unsigned int __init add_to_rb(struct printk_ringbuffer *rb, argument 1108 if (!prb_reserve(&e, rb, &dest_r)) 2387 #define prb_read_valid(rb, seq, r) false 2388 #define prb_first_valid_seq(rb) 0 2389 #define prb_next_seq(rb) 0
|
/linux-master/net/packet/ |
H A D | af_packet.c | 195 struct packet_ring_buffer *rb, 515 const struct packet_ring_buffer *rb, 522 pg_vec_pos = position / rb->frames_per_block; 523 frame_offset = position % rb->frames_per_block; 525 h.raw = rb->pg_vec[pg_vec_pos].buffer + 526 (frame_offset * rb->frame_size); 535 struct packet_ring_buffer *rb, 538 return packet_lookup_frame(po, rb, rb->head, status); 614 struct packet_ring_buffer *rb, 514 packet_lookup_frame(const struct packet_sock *po, const struct packet_ring_buffer *rb, unsigned int position, int status) argument 534 packet_current_frame(struct packet_sock *po, struct packet_ring_buffer *rb, int status) argument 613 init_prb_bdqc(struct packet_sock *po, struct packet_ring_buffer *rb, struct pgv *pg_vec, union tpacket_req_u *req_u) argument 1133 prb_lookup_block(const struct packet_sock *po, const struct packet_ring_buffer *rb, unsigned int idx, int status) argument 1146 prb_previous_blk_num(struct packet_ring_buffer *rb) argument 1157 __prb_previous_block(struct packet_sock *po, struct packet_ring_buffer *rb, int status) argument 1165 packet_previous_rx_frame(struct packet_sock *po, struct packet_ring_buffer *rb, int status) argument 1175 packet_increment_rx_head(struct packet_sock *po, struct packet_ring_buffer *rb) argument 1190 packet_previous_frame(struct packet_sock *po, struct packet_ring_buffer *rb, int status) argument 1203 packet_inc_pending(struct packet_ring_buffer *rb) argument 1208 packet_dec_pending(struct packet_ring_buffer *rb) argument 1213 packet_read_pending(const struct packet_ring_buffer *rb) argument 4393 struct packet_ring_buffer *rb; local 4559 struct packet_ring_buffer *rb; local [all...] |
/linux-master/fs/ubifs/ |
H A D | ubifs.h | 282 * @rb: rb-tree node 287 struct rb_node rb; member in struct:ubifs_old_idx 717 * @rb: link in the tree of all buds 725 struct rb_node rb; member in struct:ubifs_bud 923 * @rb: rb-tree node of rb-tree of orphans sorted by inode number 936 struct rb_node rb; member in struct:ubifs_orphan 1165 * @orph_tree: rb [all...] |
H A D | tnc.c | 56 o = rb_entry(parent, struct ubifs_old_idx, rb); 71 rb_link_node(&old_idx->rb, parent, p); 72 rb_insert_color(&old_idx->rb, &c->old_idx); 184 rbtree_postorder_for_each_entry_safe(old_idx, n, &c->old_idx, rb)
|
H A D | super.c | 922 rbtree_postorder_for_each_entry_safe(bud, n, &c->buds, rb) {
|
H A D | debug.c | 615 struct rb_node *rb; local 657 for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { 658 bud = rb_entry(rb, struct ubifs_bud, rb); 683 struct rb_node *rb; local 745 for (rb = rb_first((struct rb_root *)&c->buds); rb; r 1784 struct rb_node rb; member in struct:fsck_inode [all...] |
/linux-master/fs/kernfs/ |
H A D | dir.c | 32 #define rb_to_kn(X) rb_entry((X), struct kernfs_node, rb) 78 struct kernfs_root *ra = kernfs_root(a), *rb = kernfs_root(b); local 80 if (ra != rb) 84 db = kernfs_depth(rb->kn, b); 374 node = &pos->rb.rb_left; 376 node = &pos->rb.rb_right; 382 rb_link_node(&kn->rb, parent, node); 383 rb_insert_color(&kn->rb, &kn->parent->dir.children); 410 if (RB_EMPTY_NODE(&kn->rb)) 419 rb_erase(&kn->rb, [all...] |
/linux-master/drivers/tty/hvc/ |
H A D | hvc_iucv.c | 221 struct iucv_tty_buffer *rb; local 239 rb = list_first_entry(&priv->tty_inqueue, struct iucv_tty_buffer, list); 242 if (!rb->mbuf) { /* message not yet received ... */ 245 rb->mbuf = kmalloc(rb->msg.length, GFP_ATOMIC | GFP_DMA); 246 if (!rb->mbuf) 249 rc = __iucv_message_receive(priv->path, &rb->msg, 0, 250 rb->mbuf, rb->msg.length, NULL); 262 if (rc || (rb 904 struct iucv_tty_buffer *rb; local [all...] |