/linux-master/include/net/ |
H A D | garp.h | 82 struct rb_node node;
|
H A D | mrp.h | 95 struct rb_node node;
|
H A D | nexthop.h | 143 struct rb_node rb_node; /* entry on netns rbtree */ member in struct:nexthop
|
/linux-master/include/rdma/ |
H A D | rdmavt_qp.h | 520 struct rb_node rb_node; 519 struct rb_node rb_node; member in struct:rvt_mcast
|
/linux-master/include/xen/arm/ |
H A D | page.h | 62 if (phys_to_mach.rb_node != NULL) {
|
/linux-master/ipc/ |
H A D | mqueue.c | 62 struct rb_node rb_node; member in struct:posix_msg_tree_node 140 struct rb_node *msg_tree_rightmost; 193 struct rb_node **p, *parent = NULL; 197 p = &info->msg_tree.rb_node; 200 leaf = rb_entry(parent, struct posix_msg_tree_node, rb_node); 222 info->msg_tree_rightmost = &leaf->rb_node; 224 rb_link_node(&leaf->rb_node, parent, p); 225 rb_insert_color(&leaf->rb_node, &info->msg_tree); 236 struct rb_node *nod [all...] |
/linux-master/kernel/bpf/ |
H A D | helpers.c | 2009 * 'rb_node *', so field name of rb_node within containing struct is not 2025 struct rb_node *pos, *n; 2211 struct rb_node *n = &node_internal->rb_node; 2232 struct rb_node **link = &((struct rb_root_cached *)root)->rb_root.rb_node; 2233 struct rb_node *parent = NULL, *n = &node->rb_node;
|
H A D | local_storage.c | 71 struct rb_node *node; 76 node = root->rb_node; 106 struct rb_node **new = &(root->rb_node), *parent = NULL;
|
/linux-master/kernel/events/ |
H A D | core.c | 1634 static inline bool __group_less(struct rb_node *a, const struct rb_node *b) 1647 static inline int __group_cmp(const void *key, const struct rb_node *node) 1657 __group_cmp_ignore_cgroup(const void *key, const struct rb_node *node) 1731 struct rb_node *node; 1748 struct rb_node *next; 4237 struct rb_node *node;
|
H A D | uprobes.c | 55 struct rb_node rb_node; /* node in the rb tree */ member in struct:uprobe 631 rb_entry((node), struct uprobe, rb_node) 638 static inline int __uprobe_cmp_key(const void *key, const struct rb_node *b) 644 static inline int __uprobe_cmp(struct rb_node *a, const struct rb_node *b) 656 struct rb_node *node = rb_find(&key, &uprobes_tree, __uprobe_cmp_key); 681 struct rb_node *node; 683 node = rb_find_add(&uprobe->rb_node, &uprobes_tree, __uprobe_cmp); 926 return !RB_EMPTY_NODE(&uprobe->rb_node); [all...] |
/linux-master/kernel/locking/ |
H A D | rtmutex.c | 450 static __always_inline bool __waiter_less(struct rb_node *a, const struct rb_node *b) 499 static __always_inline bool __pi_waiter_less(struct rb_node *a, const struct rb_node *b)
|
H A D | rtmutex_common.h | 33 struct rb_node entry; 123 struct rb_node *leftmost = rb_first_cached(&lock->waiters); 130 struct rb_node *leftmost = rb_first_cached(&lock->waiters);
|
H A D | ww_mutex.h | 96 struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root); 105 struct rb_node *n = rb_next(&w->tree.entry); 114 struct rb_node *n = rb_prev(&w->tree.entry); 123 struct rb_node *n = rb_last(&lock->rtmutex.waiters.rb_root);
|
/linux-master/kernel/power/ |
H A D | swap.c | 124 struct rb_node node; 133 struct rb_node **new = &(swsusp_extents.rb_node); 134 struct rb_node *parent = NULL; 199 struct rb_node *node; 201 while ((node = swsusp_extents.rb_node)) { 216 return (swsusp_extents.rb_node != NULL);
|
H A D | wakelock.c | 29 struct rb_node node; 40 struct rb_node *node; 150 struct rb_node **node = &wakelocks_tree.rb_node; 151 struct rb_node *parent = *node;
|
/linux-master/kernel/sched/ |
H A D | core.c | 221 static inline bool rb_sched_core_less(struct rb_node *a, const struct rb_node *b) 226 static inline int rb_sched_core_cmp(const void *key, const struct rb_node *node) 279 struct rb_node *node = &p->core_node; 303 struct rb_node *node;
|
H A D | deadline.c | 89 return !RB_EMPTY_NODE(&dl_se->rb_node); 511 return rb_first_cached(&dl_rq->root) == &dl_se->rb_node; 580 static inline bool __pushable_less(struct rb_node *a, const struct rb_node *b) 596 struct rb_node *leftmost; 616 struct rb_node *leftmost; 1530 rb_entry((node), struct sched_dl_entity, rb_node) 1561 struct rb_node *leftmost = rb_first_cached(&dl_rq->root); 1597 static inline bool __dl_less(struct rb_node *a, const struct rb_node * [all...] |
H A D | fair.c | 786 static inline bool __entity_less(struct rb_node *a, const struct rb_node *b) 793 static inline void __min_vruntime_update(struct sched_entity *se, struct rb_node *node) 808 struct rb_node *node = &se->run_node; 840 struct rb_node *root = cfs_rq->tasks_timeline.rb_root.rb_node; 850 struct rb_node *left = rb_first_cached(&cfs_rq->tasks_timeline); 879 struct rb_node *node = cfs_rq->tasks_timeline.rb_root.rb_node; 909 struct rb_node *lef [all...] |
/linux-master/kernel/trace/ |
H A D | trace_stat.c | 27 struct rb_node node; 77 struct rb_node **new = &(root->rb_node), *parent = NULL; 174 struct rb_node *node; 198 struct rb_node *node = p;
|
/linux-master/lib/ |
H A D | rbtree.c | 59 static inline void rb_set_black(struct rb_node *rb) 64 static inline struct rb_node *rb_red_parent(struct rb_node *red) 66 return (struct rb_node *)red->__rb_parent_color; 75 __rb_rotate_set_parents(struct rb_node *old, struct rb_node *new, 78 struct rb_node *parent = rb_parent(old); 85 __rb_insert(struct rb_node *node, struct rb_root *root, 86 void (*augment_rotate)(struct rb_node *old, struct rb_node *ne [all...] |
H A D | rbtree_test.c | 20 struct rb_node rb; 34 struct rb_node **new = &root->rb_root.rb_node, *parent = NULL; 51 struct rb_node **new = &root->rb_root.rb_node, *parent = NULL; 88 struct rb_node **new = &root->rb_root.rb_node, *rb_parent = NULL; 112 struct rb_node **new = &root->rb_root.rb_node, *rb_parent = NULL; 158 static bool is_red(struct rb_node *r [all...] |
H A D | timerqueue.c | 20 static inline bool __timerqueue_less(struct rb_node *a, const struct rb_node *b) 75 struct rb_node *next;
|
/linux-master/mm/ |
H A D | backing-dev.c | 1037 static struct rb_node **bdi_lookup_rb_node(u64 id, struct rb_node **parentp) 1039 struct rb_node **p = &bdi_tree.rb_node; 1040 struct rb_node *parent = NULL; 1047 bdi = rb_entry(parent, struct backing_dev_info, rb_node); 1072 struct rb_node **p; 1077 bdi = rb_entry(*p, struct backing_dev_info, rb_node); 1088 struct rb_node *parent, **p; 1109 rb_link_node(&bdi->rb_node, paren [all...] |
H A D | interval_tree.c | 32 struct rb_node **link;
|
H A D | kmemleak.c | 137 * rb_node are already protected by the corresponding locks or mutex (see 146 struct rb_node rb_node; member in struct:kmemleak_object 413 struct rb_node *rb = object_tree(objflags)->rb_node; 420 object = rb_entry(rb, struct kmemleak_object, rb_node); 424 rb = object->rb_node.rb_left; 426 rb = object->rb_node.rb_right; 585 rb_erase(&object->rb_node, object_tree(object->flags)); 691 struct rb_node **lin [all...] |