/linux-master/tools/testing/selftests/sync/ |
H A D | sync_test.c | 33 #include <sys/wait.h>
|
/linux-master/tools/testing/selftests/pidfd/ |
H A D | pidfd_poll_test.c | 13 #include <sys/wait.h> 104 "death notification wait timeout\n");
|
/linux-master/tools/testing/selftests/x86/ |
H A D | test_mremap_vdso.c | 21 #include <sys/wait.h>
|
/linux-master/include/linux/ |
H A D | sysctl.h | 27 #include <linux/wait.h> 119 wait_queue_head_t wait; member in struct:ctl_table_poll 129 .wait = __WAIT_QUEUE_HEAD_INITIALIZER(name.wait) }
|
/linux-master/drivers/accessibility/speakup/ |
H A D | speakup_soft.c | 228 DEFINE_WAIT(wait); 236 prepare_to_wait(&speakup_event, &wait, TASK_INTERRUPTIBLE); 245 finish_wait(&speakup_event, &wait); 249 finish_wait(&speakup_event, &wait); 255 finish_wait(&speakup_event, &wait); 352 static __poll_t softsynth_poll(struct file *fp, struct poll_table_struct *wait) argument 357 poll_wait(fp, &speakup_event, wait);
|
/linux-master/drivers/s390/char/ |
H A D | fs3270.c | 41 wait_queue_head_t wait; /* Init & attention wait queue. */ member in struct:fs3270 69 rq->callback_data = &fp->wait; 74 rc = wait_event_interruptible(fp->wait, 81 /* Started successfully. Now wait for completion. */ 82 wait_event(fp->wait, raw3270_request_final(rq)); 97 wake_up(&fp->wait); 111 wake_up(&fp->wait); 184 wake_up(&fp->wait); 223 wake_up(&fp->wait); [all...] |
/linux-master/drivers/gpu/host1x/hw/ |
H A D | channel_hw.c | 126 if (cmd->wait.relative) 127 threshold = job_syncpt_base + cmd->wait.threshold; 129 threshold = cmd->wait.threshold; 131 submit_wait(job, cmd->wait.id, threshold, cmd->wait.next_class); 253 * Force serialization by inserting a host wait for the
|
/linux-master/drivers/md/bcache/ |
H A D | btree.h | 69 * to it, so a closure_sync() later can be used to wait for the write to 214 wait_queue_entry_t wait; member in struct:btree_op 237 wait_queue_head_t wait; member in struct:btree_check_state 244 init_wait(&op->wait); 269 int level, bool wait, 365 finish_wait(&(c)->btree_cache_wait, &(op)->wait); \
|
/linux-master/fs/jfs/ |
H A D | inode.c | 73 int jfs_commit_inode(struct inode *inode, int wait) argument 108 rc = txCommit(tid, 1, &inode, wait ? COMMIT_SYNC : 0); 117 int wait = wbc->sync_mode == WB_SYNC_ALL; local 128 jfs_flush_journal(JFS_SBI(inode->i_sb)->log, wait); 132 if (jfs_commit_inode(inode, wait)) {
|
/linux-master/fs/ |
H A D | signalfd.c | 51 static __poll_t signalfd_poll(struct file *file, poll_table *wait) argument 56 poll_wait(file, ¤t->sighand->signalfd_wqh, wait); 159 DECLARE_WAITQUEUE(wait, current); 174 add_wait_queue(¤t->sighand->signalfd_wqh, &wait); 190 remove_wait_queue(¤t->sighand->signalfd_wqh, &wait);
|
H A D | aio.c | 154 wait_queue_head_t wait; member in struct:kioctx::__anon1941 187 struct wait_queue_entry wait; member in struct:poll_iocb 787 init_waitqueue_head(&ctx->wait); 858 struct ctx_rq_wait *wait) 874 wake_up_all(&ctx->wait); 888 ctx->rq_wait = wait; 898 * There may be outstanding kiocbs, but free_ioctx() will explicitly wait on 904 struct ctx_rq_wait wait; local 910 atomic_set(&wait.count, table->nr); 911 init_completion(&wait 857 kill_ioctx(struct mm_struct *mm, struct kioctx *ctx, struct ctx_rq_wait *wait) argument 1464 struct ctx_rq_wait wait; local 1795 aio_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync, void *key) argument [all...] |
/linux-master/net/vmw_vsock/ |
H A D | af_vsock.c | 110 #include <linux/wait.h> 130 /* The default peer timeout indicates how long we will wait for a peer response 1023 poll_table *wait) 1032 poll_wait(file, sk_sleep(sk), wait); 1355 DEFINE_WAIT(wait); 1374 * will be set to zero also). Otherwise, we will either wait 1438 * enter the connected state. Here we wait for the connection to be 1442 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); local 1485 prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); local 1497 finish_wait(sk_sleep(sk), &wait); local 1022 vsock_poll(struct file *file, struct socket *sock, poll_table *wait) argument 1532 prepare_to_wait(sk_sleep(listener), &wait, TASK_INTERRUPTIBLE); local 1538 finish_wait(sk_sleep(listener), &wait); local 1549 prepare_to_wait(sk_sleep(listener), &wait, TASK_INTERRUPTIBLE); local 1551 finish_wait(sk_sleep(listener), &wait); local 1884 add_wait_queue(sk_sleep(sk), &wait); local 1893 remove_wait_queue(sk_sleep(sk), &wait); local 1899 remove_wait_queue(sk_sleep(sk), &wait); local 1908 remove_wait_queue(sk_sleep(sk), &wait); local 1912 remove_wait_queue(sk_sleep(sk), &wait); local 1916 remove_wait_queue(sk_sleep(sk), &wait); local 1980 vsock_connectible_wait_data(struct sock *sk, struct wait_queue_entry *wait, long timeout, struct vsock_transport_recv_notify_data *recv_data, size_t target) argument 1996 prepare_to_wait(sk_sleep(sk), wait, TASK_INTERRUPTIBLE); local 2032 finish_wait(sk_sleep(sk), wait); local [all...] |
/linux-master/io_uring/ |
H A D | rw.c | 693 static int io_async_buf_func(struct wait_queue_entry *wait, unsigned mode, argument 697 struct io_kiocb *req = wait->private; 701 wpq = container_of(wait, struct wait_page_queue, wait); 707 list_del_init(&wait->entry); 727 struct wait_page_queue *wait = &io->wpq; local 747 wait->wait.func = io_async_buf_func; 748 wait->wait [all...] |
H A D | poll.c | 56 static int io_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync, 178 INIT_LIST_HEAD(&poll->wait.entry); 179 init_waitqueue_func_entry(&poll->wait, io_poll_wake); 188 list_del_init(&poll->wait.entry); 205 * wake_up_pollfree() will wait for us. However, taking the waitqueue 417 list_del_init(&poll->wait.entry); 429 static int io_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync, argument 432 struct io_kiocb *req = wqe_to_req(wait); 433 struct io_poll *poll = container_of(wait, struct io_poll, wait); [all...] |
/linux-master/drivers/iio/ |
H A D | industrialio-event.c | 19 #include <linux/wait.h> 28 * @wait: wait queue to allow blocking reads of events 37 wait_queue_head_t wait; member in struct:iio_event_interface 84 wake_up_poll(&ev_int->wait, EPOLLIN); 94 * @wait: Poll table pointer to add the wait queue on 100 struct poll_table_struct *wait) 110 poll_wait(filep, &ev_int->wait, wait); 99 iio_event_poll(struct file *filep, struct poll_table_struct *wait) argument [all...] |
/linux-master/drivers/tty/ |
H A D | n_hdlc.c | 306 /* wait up sleeping writers */ 434 DECLARE_WAITQUEUE(wait, current); 441 add_wait_queue(&tty->read_wait, &wait); 471 remove_wait_queue(&tty->read_wait, &wait); 524 DECLARE_WAITQUEUE(wait, current); 537 add_wait_queue(&tty->write_wait, &wait); 559 remove_wait_queue(&tty->write_wait, &wait); 642 * @wait: wait queue for operations 649 poll_table *wait) 648 n_hdlc_tty_poll(struct tty_struct *tty, struct file *filp, poll_table *wait) argument [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | sdma.c | 45 /* max wait time for a SDMA engine to indicate it has halted */ 260 * sdma_wait_for_packet_egress() - wait for the VL FIFO occupancy for 296 * sdma_wait() - wait for packet egress to complete for all SDMA engines, 327 struct iowait *wait = tx->wait; local 340 if (iowait_sdma_dec(wait)) 341 iowait_drain_wakeup(wait); 1725 struct iowait *wait, *nw, *twait; local 1742 wait, 1748 if (!wait 2312 sdma_check_progress( struct sdma_engine *sde, struct iowait_work *wait, struct sdma_txreq *tx, bool pkts_sent) argument 2353 sdma_send_txreq(struct sdma_engine *sde, struct iowait_work *wait, struct sdma_txreq *tx, bool pkts_sent) argument 2432 sdma_send_txlist(struct sdma_engine *sde, struct iowait_work *wait, struct list_head *tx_list, u16 *count_out) argument [all...] |
/linux-master/arch/mips/include/asm/octeon/ |
H A D | cvmx-pow.h | 260 uint64_t wait:1; member in struct:__anon29::__anon30 265 uint64_t wait:1; 481 * case can happen, regardless of the wait bit value 527 * the wait bit was set. 609 * the wait bit was set. 1234 uint64_t wait:1; member in struct:__anon53::__anon54 1238 uint64_t wait:1; 1342 * This function does NOT wait for previous tag switches to complete, 1345 * @wait: When set, call stalls until work becomes available, or times out. 1352 wait) 1351 cvmx_pow_work_request_sync_nocheck(cvmx_pow_wait_t wait) argument 1385 cvmx_pow_work_request_sync(cvmx_pow_wait_t wait) argument 1438 cvmx_pow_work_request_async_nocheck(int scr_addr, cvmx_pow_wait_t wait) argument 1467 cvmx_pow_work_request_async(int scr_addr, cvmx_pow_wait_t wait) argument [all...] |
/linux-master/drivers/media/i2c/ |
H A D | saa7110.c | 20 #include <linux/wait.h> 179 DEFINE_WAIT(wait); 186 prepare_to_wait(&decoder->wq, &wait, TASK_UNINTERRUPTIBLE); 188 finish_wait(&decoder->wq, &wait); 221 prepare_to_wait(&decoder->wq, &wait, TASK_UNINTERRUPTIBLE); 223 finish_wait(&decoder->wq, &wait);
|
/linux-master/drivers/spi/ |
H A D | spi-sh.c | 77 wait_queue_head_t wait; member in struct:spi_sh_data 187 ret = wait_event_interruptible_timeout(ss->wait, 203 ret = wait_event_interruptible_timeout(ss->wait, 239 ret = wait_event_interruptible_timeout(ss->wait, 374 wake_up(&ss->wait); 433 init_waitqueue_head(&ss->wait);
|
/linux-master/fs/crypto/ |
H A D | fname.c | 96 DECLARE_CRYPTO_WAIT(wait); 121 crypto_req_done, &wait); 126 res = crypto_wait_req(crypto_skcipher_encrypt(req), &wait); 152 DECLARE_CRYPTO_WAIT(wait); 165 crypto_req_done, &wait); 174 res = crypto_wait_req(crypto_skcipher_decrypt(req), &wait);
|
/linux-master/sound/core/seq/ |
H A D | seq_memory.c | 275 wait_queue_entry_t wait; local 282 init_waitqueue_entry(&wait, current); 292 add_wait_queue(&pool->output_sleep, &wait); 300 remove_wait_queue(&pool->output_sleep, &wait); 425 /* poll wait */ 427 poll_table *wait) 429 poll_wait(file, &pool->output_sleep, wait); 490 /* wait for closing all threads */ 426 snd_seq_pool_poll_wait(struct snd_seq_pool *pool, struct file *file, poll_table *wait) argument
|
/linux-master/sound/core/ |
H A D | hwdep.c | 73 wait_queue_entry_t wait; local 93 init_waitqueue_entry(&wait, current); 94 add_wait_queue(&hw->open_wait, &wait); 128 remove_wait_queue(&hw->open_wait, &wait); 165 static __poll_t snd_hwdep_poll(struct file * file, poll_table * wait) argument 169 return hw->ops.poll(hw, file, wait);
|
/linux-master/fs/xfs/ |
H A D | xfs_extent_busy.c | 340 * value can be used to wait for at least one of the currently busy extents 587 * to wait for committed busy extents to resolve. If we are being told just to 604 DEFINE_WAIT (wait); 625 prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); 631 finish_wait(&pag->pagb_wait, &wait); 640 DEFINE_WAIT (wait); 645 prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); 650 finish_wait(&pag->pagb_wait, &wait);
|
/linux-master/drivers/hid/usbhid/ |
H A D | hiddev.c | 159 wake_up_interruptible(&hiddev->wait); 321 DEFINE_WAIT(wait); 339 prepare_to_wait(&list->hiddev->wait, &wait, TASK_INTERRUPTIBLE); 359 finish_wait(&list->hiddev->wait, &wait); 364 finish_wait(&list->hiddev->wait, &wait); 412 static __poll_t hiddev_poll(struct file *file, poll_table *wait) argument 416 poll_wait(file, &list->hiddev->wait, wai [all...] |