/linux-master/include/trace/events/ |
H A D | rcu.h | 514 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen), 516 TP_ARGS(rcuname, rhp, qlen), 520 __field(void *, rhp) 527 __entry->rhp = rhp; 528 __entry->func = rhp->func; 532 TP_printk("%s rhp=%p func=%ps %ld", 533 __entry->rcuname, __entry->rhp, __entry->func, 573 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset, 576 TP_ARGS(rcuname, rhp, offse [all...] |
/linux-master/kernel/rcu/ |
H A D | rcu_segcblist.c | 28 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp) argument 30 *rclp->tail = rhp; 31 rclp->tail = &rhp->next; 37 * any contents of the first. If rhp is non-NULL, enqueue it as the sole 45 struct rcu_head *rhp) 53 if (!rhp) { 56 rhp->next = NULL; 57 srclp->head = rhp; 58 srclp->tail = &rhp->next; 69 struct rcu_head *rhp; local 43 rcu_cblist_flush_enqueue(struct rcu_cblist *drclp, struct rcu_cblist *srclp, struct rcu_head *rhp) argument 340 rcu_segcblist_enqueue(struct rcu_segcblist *rsclp, struct rcu_head *rhp) argument 360 rcu_segcblist_entrain(struct rcu_segcblist *rsclp, struct rcu_head *rhp) argument [all...] |
H A D | srcutiny.c | 116 struct rcu_head *rhp; local 139 rhp = lh; 141 debug_rcu_head_callback(rhp); 143 rhp->func(rhp); 179 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, argument 184 rhp->func = func; 185 rhp->next = NULL; 187 *ssp->srcu_cb_tail = rhp; 188 ssp->srcu_cb_tail = &rhp [all...] |
H A D | rcu_segcblist.h | 24 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp); 27 struct rcu_head *rhp); 139 struct rcu_head *rhp); 141 struct rcu_head *rhp);
|
H A D | sync.c | 27 static void rcu_sync_func(struct rcu_head *rhp); 36 * @rhp: Pointer to rcu_head in rcu_sync structure to use for synchronization 57 static void rcu_sync_func(struct rcu_head *rhp) argument 59 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head);
|
H A D | update.c | 516 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, argument 520 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); 524 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ 610 struct early_boot_kfree_rcu *rhp; local 617 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); 618 if (!WARN_ON_ONCE(!rhp)) 619 kfree_rcu(rhp, rh);
|
H A D | rcu.h | 252 static inline void debug_rcu_head_callback(struct rcu_head *rhp) argument 254 if (unlikely(!rhp->func)) 255 kmem_dump_obj(rhp); 563 struct rcu_head *rhp, 577 struct rcu_head *rhp, 582 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \
|
H A D | tasks.h | 321 static void call_rcu_tasks_generic(struct rcu_head *rhp, rcu_callback_t func, argument 333 rhp->next = NULL; 334 rhp->func = func; 364 rcu_segcblist_enqueue(&rtpcp->cblist, rhp); 383 static void rcu_barrier_tasks_generic_cb(struct rcu_head *rhp) argument 388 rtpcp = container_of(rhp, struct rcu_tasks_percpu, barrier_q_head); 516 struct rcu_head *rhp; local 541 for (rhp = rcu_cblist_dequeue(&rcl); rhp; rhp 1107 call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) argument 1293 call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_t func) argument 1924 call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) argument 2052 test_rcu_tasks_callback(struct rcu_head *rhp) argument [all...] |
H A D | srcutree.c | 1202 static void srcu_leak_callback(struct rcu_head *rhp) argument 1210 struct rcu_head *rhp, bool do_norm) 1234 if (rhp) 1235 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp); 1273 if (rhp) { 1337 static void __call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, argument 1340 if (debug_rcu_head_queue(rhp)) { 1342 WRITE_ONCE(rhp->func, srcu_leak_callback); 1346 rhp->func = func; 1347 (void)srcu_gp_start_if_needed(ssp, rhp, do_nor 1209 srcu_gp_start_if_needed(struct srcu_struct *ssp, struct rcu_head *rhp, bool do_norm) argument 1367 call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, rcu_callback_t func) argument 1555 srcu_barrier_cb(struct rcu_head *rhp) argument 1722 struct rcu_head *rhp; local [all...] |
H A D | rcuscale.c | 442 static void rcu_scale_async_cb(struct rcu_head *rhp) argument 445 kfree(rhp); 458 struct rcu_head *rhp = NULL; local 503 if (!rhp) 504 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); 505 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { 507 cur_ops->async(rhp, rcu_scale_async_cb); 508 rhp = NULL; 513 kfree(rhp); /* Becaus [all...] |
H A D | rcutorture.c | 1712 static void rcu_torture_timer_cb(struct rcu_head *rhp) argument 1714 kfree(rhp); 2065 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); local 2067 if (rhp) 2068 cur_ops->call(rhp, rcu_torture_timer_cb); 2303 struct rcu_head *rhp; local 2310 rhp = kmem_cache_alloc(kcp, GFP_KERNEL); 2311 if (WARN_ON_ONCE(!rhp)) { 2315 pr_alert("mem_dump_obj() slab test: rcu_torture_stats = %px, &rhp 2537 rcu_torture_fwd_prog_cb(struct rcu_head *rhp) argument 2609 rcu_torture_fwd_cb_cr(struct rcu_head *rhp) argument 3045 struct rcu_head *rhp = rcu_void; local 3451 rcu_torture_leak_cb(struct rcu_head *rhp) argument 3455 rcu_torture_err_cb(struct rcu_head *rhp) argument 3480 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); local [all...] |
H A D | tree_nocb.h | 320 * Flush the ->nocb_bypass queue into ->cblist, enqueuing rhp if non-NULL. 328 * Note that this function always returns true if rhp is NULL. 334 struct rcu_head *rhp = rhp_in; local 339 if (rhp && !rcu_cblist_n_cbs(&rdp->nocb_bypass)) { 344 if (rhp) 353 if (lazy && rhp) { 354 rcu_cblist_enqueue(&rdp->nocb_bypass, rhp); 355 rhp = NULL; 357 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); 367 * Flush the ->nocb_bypass queue into ->cblist, enqueuing rhp i 374 rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, unsigned long j, bool lazy) argument 415 rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp, bool *was_alldone, unsigned long flags, bool lazy) argument 1774 rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, unsigned long j, bool lazy) argument [all...] |
H A D | tiny.c | 162 static void tiny_rcu_leak_callback(struct rcu_head *rhp) argument
|
H A D | tree.h | 467 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | mem.c | 388 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); 391 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, argument 397 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, 410 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, 419 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, 432 struct c4iw_dev *rhp; local 440 rhp = php->rhp; 458 mhp->rhp = rhp; 498 struct c4iw_dev *rhp; local 601 struct c4iw_dev *rhp; local 702 struct c4iw_dev *rhp; local 729 c4iw_invalidate_mr(struct c4iw_dev *rhp, u32 rkey) argument [all...] |
H A D | provider.c | 65 struct c4iw_dev *rhp; local 69 rhp = to_c4iw_dev(ucontext->ibucontext.device); 73 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); 81 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); local 87 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); 93 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; 114 mm->addr = virt_to_phys(rhp->rdev.status_page); 196 struct c4iw_dev *rhp; local 200 rhp = php->rhp; 214 struct c4iw_dev *rhp; local [all...] |
H A D | qp.c | 720 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); 744 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); 807 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); 916 xa_lock_irqsave(&qhp->rhp->qps, flags); 918 if (qhp->rhp->db_state == NORMAL) 921 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); 925 xa_unlock_irqrestore(&qhp->rhp->qps, flags); 933 xa_lock_irqsave(&qhp->rhp->qps, flags); 935 if (qhp->rhp->db_state == NORMAL) 938 add_to_fc_list(&qhp->rhp 1085 struct c4iw_dev *rhp; local 1692 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) argument 1749 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) argument 1832 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) argument 2072 struct c4iw_dev *rhp; local 2113 struct c4iw_dev *rhp; local 2359 struct c4iw_dev *rhp; local 2666 struct c4iw_dev *rhp; local 2798 struct c4iw_dev *rhp; local [all...] |
H A D | cq.c | 351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); 803 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); 844 c4iw_invalidate_mr(qhp->rhp, 933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); 984 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); 990 destroy_cq(&chp->rhp->rdev, &chp->cq, 1003 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); local 1020 if (vector >= rhp->rdev.lldi.nciq) 1056 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); 1078 ret = create_cq(&rhp [all...] |
H A D | iw_cxgb4.h | 343 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) argument 345 return xa_load(&rhp->cqs, cqid); 348 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) argument 350 return xa_load(&rhp->qps, qpid); 363 struct c4iw_dev *rhp; member in struct:c4iw_pd 392 struct c4iw_dev *rhp; member in struct:c4iw_mr 410 struct c4iw_dev *rhp; member in struct:c4iw_mw 424 struct c4iw_dev *rhp; member in struct:c4iw_cq 481 struct c4iw_dev *rhp; member in struct:c4iw_qp 504 struct c4iw_dev *rhp; member in struct:c4iw_srq [all...] |
H A D | ev.c | 100 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE,
|
/linux-master/include/linux/ |
H A D | rcupdate_trace.h | 87 void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func); 96 static inline void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) { BUG(); } argument
|
H A D | rcupdate.h | 1052 * @rhp: The rcu_head structure to initialize. 1060 static inline void rcu_head_init(struct rcu_head *rhp) argument 1062 rhp->func = (rcu_callback_t)~0L; 1067 * @rhp: The rcu_head structure to test. 1068 * @f: The function passed to call_rcu() along with @rhp. 1070 * Returns @true if the @rhp has been passed to call_rcu() with @func, 1072 * the case where @rhp has already been invoked after a grace period. 1076 * of the pointer to the structure containing @rhp. 1079 rcu_head_after_call_rcu(struct rcu_head *rhp, rcu_callback_t f) argument 1081 rcu_callback_t func = READ_ONCE(rhp [all...] |
/linux-master/drivers/media/usb/pvrusb2/ |
H A D | pvrusb2-v4l2.c | 42 struct pvr2_ioread *rhp; member in struct:pvr2_v4l2_fh 915 if (fhp->rhp) { 918 sp = pvr2_ioread_get_stream(fhp->rhp); 920 pvr2_ioread_destroy(fhp->rhp); 921 fhp->rhp = NULL; 1048 if (fh->rhp) return 0; 1064 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); 1065 if (!fh->rhp) { 1075 return pvr2_ioread_set_enabled(fh->rhp,!0); 1118 if (!fh->rhp) { [all...] |
/linux-master/include/linux/sched/ |
H A D | mm.h | 62 static inline void __mmdrop_delayed(struct rcu_head *rhp) argument 64 struct mm_struct *mm = container_of(rhp, struct mm_struct, delayed_drop);
|
H A D | task.h | 123 extern void __put_task_struct_rcu_cb(struct rcu_head *rhp);
|