/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | icp_qat_hw_20_comp.h | 69 __u16 lazy; member in struct:icp_qat_hw_comp_20_config_csr_upper 102 QAT_FIELD_SET(val32, csr.lazy,
|
H A D | adf_gen4_dc.c | 42 hw_comp_upper_csr.lazy = ICP_QAT_HW_COMP_20_CONFIG_CSR_LAZY_PARAM_DEFAULT_VAL;
|
/linux-master/drivers/opp/ |
H A D | opp.h | 206 struct list_head node, lazy; member in struct:opp_table 271 return unlikely(!list_empty(&opp_table->lazy));
|
H A D | of.c | 153 list_del(&opp_table->lazy); 167 bool lazy = false; local 200 lazy = true; 204 if (lazy) { 210 list_add(&opp_table->lazy, &lazy_opp_tables); 412 list_for_each_entry_safe(opp_table, temp, &lazy_opp_tables, lazy) { 413 bool lazy = false; local 437 lazy = true; 448 lazy = false; 455 /* All required opp-tables found, remove from lazy lis [all...] |
H A D | core.c | 1444 INIT_LIST_HEAD(&opp_table->lazy);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_fence.h | 94 bool lazy,
|
H A D | vmwgfx_irq.c | 163 bool lazy, 211 if (lazy) 162 vmw_fallback_wait(struct vmw_private *dev_priv, bool lazy, bool fifo_idle, uint32_t seqno, bool interruptible, unsigned long timeout) argument
|
H A D | vmwgfx_fence.c | 522 int vmw_fence_obj_wait(struct vmw_fence_obj *fence, bool lazy, argument 763 ret = vmw_fence_obj_wait(fence, arg->lazy, true, timeout);
|
H A D | vmwgfx_drv.h | 1038 bool lazy,
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_fence.h | 26 int nouveau_fence_wait(struct nouveau_fence *, bool lazy, bool intr);
|
H A D | nouveau_fence.c | 340 nouveau_fence_wait(struct nouveau_fence *fence, bool lazy, bool intr) argument 344 if (!lazy)
|
/linux-master/kernel/rcu/ |
H A D | tree_nocb.h | 262 * can elapse before lazy callbacks are flushed. Lazy callbacks 264 * however, LAZY_FLUSH_JIFFIES will ensure no lazy callbacks are 331 unsigned long j, bool lazy) 348 * If the new CB requested was a lazy one, queue it onto the main 351 * the lazy CB is ordered with the existing CBs in the bypass list. 353 if (lazy && rhp) { 375 unsigned long j, bool lazy) 381 return rcu_nocb_do_flush_bypass(rdp, rhp, j, lazy); 417 bool lazy) 469 if (rdp->nocb_nobypass_count < nocb_nobypass_lim_per_jiffy && !lazy) { 330 rcu_nocb_do_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp_in, unsigned long j, bool lazy) argument 374 rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, unsigned long j, bool lazy) argument 415 rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp, bool *was_alldone, unsigned long flags, bool lazy) argument 626 call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head, rcu_callback_t func, unsigned long flags, bool lazy) argument 696 bool lazy = false; local 1774 rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, unsigned long j, bool lazy) argument 1780 call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head, rcu_callback_t func, unsigned long flags, bool lazy) argument [all...] |
H A D | tree.h | 284 long lazy_len; /* Length of buffered lazy callbacks. */ 468 unsigned long j, bool lazy); 470 rcu_callback_t func, unsigned long flags, bool lazy);
|
H A D | tree.c | 2713 bool lazy; local 2737 lazy = lazy_in && !rcu_async_should_hurry(); 2753 call_rcu_nocb(rdp, head, func, flags, lazy); 2765 * flush all lazy callbacks (including the new one) to the main ->cblist while 2794 * By default the callbacks are 'lazy' and are kept hidden from the main 3406 * Queue a request for lazy invocation of the appropriate free routine 4021 * if it's fully lazy.
|
/linux-master/kernel/ |
H A D | irq_work.c | 110 /* If the work is "lazy", handle it from next tick if any */ 159 * IRQ_WORK_HARD_IRQ are added to the lazy list and a HARD work 186 struct llist_head *raised, *lazy; local 189 lazy = this_cpu_ptr(&lazy_list); 192 if (llist_empty(lazy))
|
/linux-master/include/uapi/drm/ |
H A D | vmwgfx_drm.h | 634 * @lazy: Set to 1 if timing is not critical. Allow more than a kernel tick 647 __s32 lazy; member in struct:drm_vmw_fence_wait_arg
|
/linux-master/mm/ |
H A D | vmalloc.c | 876 struct rb_list lazy; member in struct:vmap_node 2226 if (RB_EMPTY_ROOT(&vn->lazy.root)) 2229 spin_lock(&vn->lazy.lock); 2230 WRITE_ONCE(vn->lazy.root.rb_node, NULL); 2231 list_replace_init(&vn->lazy.head, &vn->purge_list); 2232 spin_unlock(&vn->lazy.lock); 2329 spin_lock(&vn->lazy.lock); 2330 insert_vmap_area(va, &vn->lazy.root, &vn->lazy.head); 2331 spin_unlock(&vn->lazy [all...] |
/linux-master/include/asm-generic/ |
H A D | hyperv-tlfs.h | 355 u64 lazy:1; member in struct:hv_stimer_config::__anon110
|
/linux-master/arch/arm/crypto/ |
H A D | poly1305-armv4.pl | 585 @ lazy reduction as discussed in "NEON crypto" by D.J. Bernstein 993 @ lazy reduction interleaved with base 2^32 -> base 2^26 of 1163 @ lazy reduction, but without narrowing
|
/linux-master/arch/arm64/crypto/ |
H A D | poly1305-armv8.pl | 700 // lazy reduction as discussed in "NEON crypto" by D.J. Bernstein 845 // lazy reduction, but without narrowing
|
/linux-master/arch/sparc/lib/ |
H A D | checksum_32.S | 411 addx %g5, %g0, %g5 ! I am now to lazy to optimize this (question it
|