/linux-master/include/linux/ |
H A D | regmap.h | 512 struct regmap_async *async); 542 * must serialise with respect to non-async I/O. 1242 bool *change, bool async, bool force); 1385 bool *change, bool async, bool force); 1390 bool *change, bool async, bool force); 1773 bool *change, bool async, bool force) 1802 bool *change, bool async, bool force) 1811 bool *change, bool async, bool force) 1771 regmap_update_bits_base(struct regmap *map, unsigned int reg, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument 1800 regmap_field_update_bits_base(struct regmap_field *field, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument 1808 regmap_fields_update_bits_base(struct regmap_field *field, unsigned int id, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument
|
H A D | blkdev.h | 1003 static inline void blk_flush_plug(struct blk_plug *plug, bool async) argument 1006 __blk_flush_plug(plug, async); 1040 static inline void blk_flush_plug(struct blk_plug *plug, bool async) argument
|
H A D | libata.h | 25 #include <linux/async.h>
|
/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 272 return false; /* async events don't have a request */ 1338 struct nvme_tcp_request *async = &ctrl->async_req; local 1340 page_frag_free(async->pdu); 1346 struct nvme_tcp_request *async = &ctrl->async_req; local 1349 async->pdu = page_frag_alloc(&queue->pf_cache, 1352 if (!async->pdu) 1355 async->queue = &ctrl->queues[0];
|
H A D | pci.c | 8 #include <linux/async.h>
|
/linux-master/drivers/base/regmap/ |
H A D | regmap.c | 1445 struct regmap_async *async; local 1454 async = list_first_entry_or_null(&map->async_free, 1457 list_del(&async->list); 1458 kfree(async->work_buf); 1459 kfree(async); 1693 if (map->async && map->bus && map->bus->async_write) { 1694 struct regmap_async *async; local 1699 async = list_first_entry_or_null(&map->async_free, 1702 if (async) 1703 list_del(&async 2227 regmap_field_update_bits_base(struct regmap_field *field, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument 2276 regmap_fields_update_bits_base(struct regmap_field *field, unsigned int id, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument 3231 regmap_update_bits_base(struct regmap *map, unsigned int reg, unsigned int mask, unsigned int val, bool *change, bool async, bool force) argument 3274 regmap_async_complete_cb(struct regmap_async *async, int ret) argument [all...] |
/linux-master/virt/kvm/ |
H A D | kvm_mm.h | 24 bool *async, bool write_fault, bool *writable);
|
H A D | kvm_main.c | 2825 static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, argument 2850 if (async) 2967 * @async: whether this function need to wait IO complete if the 2978 bool *async, bool write_fault, bool *writable) 2985 BUG_ON(atomic && async); 2993 npages = hva_to_pfn_slow(addr, async, write_fault, interruptible, 3002 (!async && check_user_page_hwpoison(addr))) { 3019 if (async && vma_is_valid(vma, write_fault)) 3020 *async = true; 3029 bool atomic, bool interruptible, bool *async, 2977 hva_to_pfn(unsigned long addr, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable) argument 3028 __gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool interruptible, bool *async, bool write_fault, bool *writable, hva_t *hva) argument [all...] |
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 4347 bool async; local 4384 async = false; 4385 fault->pfn = __gfn_to_pfn_memslot(slot, fault->gfn, false, false, &async, 4388 if (!async) 4579 WARN_ONCE(1, "Unexpected host async PF flags: %x\n", flags);
|
/linux-master/init/ |
H A D | main.c | 79 #include <linux/async.h> 1445 /* need to finish all async __init code before freeing the memory */
|
H A D | initramfs.c | 3 #include <linux/async.h>
|
/linux-master/fs/fuse/ |
H A D | file.c | 814 if (ia->io->async) 1008 * Congested and only async pages left, so skip the 1105 if (ia->io->async) 1601 if (!io->async || nres < 0) { 3014 * By default, we want to optimize all I/Os with async request 3017 io->async = ff->fm->fc->async_dio; 3022 if (io->async && !io->write && offset + count > i_size) { 3035 if (io->async && io->blocking) { 3052 if (io->async) { 3057 /* we have a non-extending, async reques [all...] |
H A D | fuse_i.h | 342 int async; member in struct:fuse_io_priv 359 .async = 0, \
|
/linux-master/drivers/firmware/arm_scmi/ |
H A D | raw_mode.c | 67 * - message*: used to send sync/async commands and read back immediate and 198 * @async_response: A completion to be, optionally, used for async waits: it 337 struct scmi_chan_info *cinfo, bool async) 347 if (async) { 480 /* Wait also for an async delayed response if needed */ 599 * @async: A flag stating if an asynchronous command is required. 613 bool async) 629 rw = scmi_xfer_raw_waiter_get(raw, xfer, cinfo, async); 673 * @async: A flag stating if an asynchronous command is required. 678 void *buf, size_t len, u8 chan_id, bool async) 336 scmi_xfer_raw_waiter_get(struct scmi_raw_mode_info *raw, struct scmi_xfer *xfer, struct scmi_chan_info *cinfo, bool async) argument 611 scmi_do_xfer_raw_start(struct scmi_raw_mode_info *raw, struct scmi_xfer *xfer, u8 chan_id, bool async) argument 677 scmi_raw_message_send(struct scmi_raw_mode_info *raw, void *buf, size_t len, u8 chan_id, bool async) argument 801 scmi_dbg_raw_mode_common_write(struct file *filp, const char __user *buf, size_t count, loff_t *ppos, bool async) argument [all...] |
/linux-master/drivers/ata/ |
H A D | libata-core.c | 5175 bool async) 5204 if (!async) 5209 bool async) 5230 async); 5263 bool async) 5267 async); 5328 * level. sas suspend/resume is async to allow parallel port recovery 5173 ata_port_request_pm(struct ata_port *ap, pm_message_t mesg, unsigned int action, unsigned int ehi_flags, bool async) argument 5208 ata_port_suspend(struct ata_port *ap, pm_message_t mesg, bool async) argument 5262 ata_port_resume(struct ata_port *ap, pm_message_t mesg, bool async) argument
|
/linux-master/drivers/scsi/qla2xxx/ |
H A D | qla_edif.c | 27 uint32_t async; member in union:edb_node::__anon85
|
/linux-master/drivers/scsi/hisi_sas/ |
H A D | hisi_sas_main.c | 1541 ASYNC_DOMAIN_EXCLUSIVE(async); 1552 phy, &async); 1555 async_synchronize_full_domain(&async); 1912 ASYNC_DOMAIN_EXCLUSIVE(async); 1929 device, &async); 1932 async_synchronize_full_domain(&async);
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu.h | 464 bool async; member in struct:amdgpu_flip_work
|
/linux-master/drivers/ufs/core/ |
H A D | ufshcd.c | 12 #include <linux/async.h> 10699 /* Hold auto suspend until async scan completes */
|
/linux-master/drivers/usb/dwc3/ |
H A D | gadget.c | 280 static int __dwc3_gadget_wakeup(struct dwc3 *dwc, bool async); 2309 static int __dwc3_gadget_wakeup(struct dwc3 *dwc, bool async) argument 2340 if (async) 2362 if (async)
|
/linux-master/drivers/scsi/ |
H A D | scsi_scan.c | 36 #include <linux/async.h> 95 #define SCSI_SCAN_TYPE_DEFAULT "async" 104 MODULE_PARM_DESC(scan, "sync, async, manual, or none. " 126 * scsi_enable_async_suspend - Enable async suspend and resume 131 * If a user has disabled async probing a likely reason is due to a 135 if (strncmp(scsi_scan_type, "async", 5) != 0) 157 * sleep a little. Even if we never get memory, the async 863 * @async: 1 if this device is being scanned asynchronously 874 blist_flags_t *bflags, int async) 1115 if (!async 873 scsi_add_lun(struct scsi_device *sdev, unsigned char *inq_result, blist_flags_t *bflags, int async) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_vma.c | 137 * use an async variant of intel_gt_pm_put(). 512 /* Wait for or await async unbinds touching our range */ 551 * Also note that we do not want to track the async vma as 1779 /* Wait for async active retire */ 2013 struct dma_fence *__i915_vma_evict(struct i915_vma *vma, bool async) argument 2051 /* Object backend must be async capable. */ 2052 GEM_WARN_ON(async && !vma->resource->bi.pages_rsgt); 2061 if (async) 2074 if (!async) { 2228 * async unbin [all...] |
/linux-master/block/ |
H A D | blk-mq.c | 2231 * @async: If we want to run the queue asynchronously. 2237 void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async) argument 2244 WARN_ON_ONCE(!async && in_interrupt()); 2246 might_sleep_if(!async && hctx->flags & BLK_MQ_F_BLOCKING); 2263 if (async || !cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)) { 2297 * @async: If we want to run the queue asynchronously. 2299 void blk_mq_run_hw_queues(struct request_queue *q, bool async) argument 2317 blk_mq_run_hw_queue(hctx, async); 2412 void blk_mq_start_stopped_hw_queue(struct blk_mq_hw_ctx *hctx, bool async) argument 2418 blk_mq_run_hw_queue(hctx, async); 2422 blk_mq_start_stopped_hw_queues(struct request_queue *q, bool async) argument [all...] |
/linux-master/net/tls/ |
H A D | tls_sw.c | 54 bool async; 271 if (darg->async) { 296 /* all completions have run, we're not doing async anymore */ 297 darg->async = false; 302 darg->async = false; 1448 * async | Async decrypt allowed | Async crypto used / in progress 1609 if (unlikely(darg->async)) { 1648 /* keep going even for ->async, the code below is TLS 1.3 */ 1693 darg->async = false; 1967 bool async local [all...] |
/linux-master/drivers/usb/host/ |
H A D | sl811-hcd.c | 299 * from the previous frame), and the rest of the time is async 316 else if (!list_empty(&sl811->async)) 317 ep = container_of(sl811->async.next, 331 if (ep->schedule.next == &sl811->async) 441 /* async deschedule? */ 469 /* we might turn SOFs back on again for the async schedule */ 746 if (sl811->periodic_count == 0 && list_empty(&sl811->async)) 898 list_add_tail(&ep->schedule, &sl811->async); 1436 list_for_each_entry (ep, &sl811->async, schedule) { 1459 if (!list_empty(&sl811->async)) [all...] |