Lines Matching refs:queue
25 int wg_packet_queue_init(struct crypt_queue *queue, work_func_t function,
30 memset(queue, 0, sizeof(*queue));
31 queue->last_cpu = -1;
32 ret = ptr_ring_init(&queue->ring, len, GFP_KERNEL);
35 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue);
36 if (!queue->worker) {
37 ptr_ring_cleanup(&queue->ring, NULL);
43 void wg_packet_queue_free(struct crypt_queue *queue, bool purge)
45 free_percpu(queue->worker);
46 WARN_ON(!purge && !__ptr_ring_empty(&queue->ring));
47 ptr_ring_cleanup(&queue->ring, purge ? __skb_array_destroy_skb : NULL);
51 #define STUB(queue) ((struct sk_buff *)&queue->empty)
53 void wg_prev_queue_init(struct prev_queue *queue)
55 NEXT(STUB(queue)) = NULL;
56 queue->head = queue->tail = STUB(queue);
57 queue->peeked = NULL;
58 atomic_set(&queue->count, 0);
66 static void __wg_prev_queue_enqueue(struct prev_queue *queue, struct sk_buff *skb)
69 WRITE_ONCE(NEXT(xchg_release(&queue->head, skb)), skb);
72 bool wg_prev_queue_enqueue(struct prev_queue *queue, struct sk_buff *skb)
74 if (!atomic_add_unless(&queue->count, 1, MAX_QUEUED_PACKETS))
76 __wg_prev_queue_enqueue(queue, skb);
80 struct sk_buff *wg_prev_queue_dequeue(struct prev_queue *queue)
82 struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail));
84 if (tail == STUB(queue)) {
87 queue->tail = next;
92 queue->tail = next;
93 atomic_dec(&queue->count);
96 if (tail != READ_ONCE(queue->head))
98 __wg_prev_queue_enqueue(queue, STUB(queue));
101 queue->tail = next;
102 atomic_dec(&queue->count);