Lines Matching refs:rbwork
766 struct rb_irq_work *rbwork = container_of(work, struct rb_irq_work, work);
769 (void)atomic_fetch_inc_release(&rbwork->seq);
771 wake_up_all(&rbwork->waiters);
772 if (rbwork->full_waiters_pending || rbwork->wakeup_full) {
775 container_of(rbwork, struct ring_buffer_per_cpu, irq_work);
779 rbwork->wakeup_full = false;
780 rbwork->full_waiters_pending = false;
786 wake_up_all(&rbwork->full_waiters);
801 struct rb_irq_work *rbwork;
812 rbwork = &buffer->irq_work;
823 rbwork = &cpu_buffer->irq_work;
827 irq_work_queue(&rbwork->work);
862 rb_wait_cond(struct rb_irq_work *rbwork, struct trace_buffer *buffer,
892 rbwork->full_waiters_pending = true;
894 rbwork->waiters_pending = true;
911 struct rb_irq_work *rbwork = rdata->irq_work;
913 return atomic_read_acquire(&rbwork->seq) != rdata->seq;
933 struct rb_irq_work *rbwork;
943 rbwork = &buffer->irq_work;
950 rbwork = &cpu_buffer->irq_work;
954 waitq = &rbwork->full_waiters;
956 waitq = &rbwork->waiters;
961 rdata.irq_work = rbwork;
962 rdata.seq = atomic_read_acquire(&rbwork->seq);
967 rb_wait_cond(rbwork, buffer, cpu, full, cond, data));
991 struct rb_irq_work *rbwork;
994 rbwork = &buffer->irq_work;
1001 rbwork = &cpu_buffer->irq_work;
1005 poll_wait(filp, &rbwork->full_waiters, poll_table);
1022 rbwork->full_waiters_pending = true;
1026 poll_wait(filp, &rbwork->waiters, poll_table);
1027 rbwork->waiters_pending = true;