Lines Matching refs:TIF_PATCH_PENDING
132 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING));
139 WARN_ON_ONCE(test_tsk_thread_flag(task, TIF_PATCH_PENDING));
196 * 1) Enforce the order of the TIF_PATCH_PENDING read and the
200 * 2) Enforce the order of the TIF_PATCH_PENDING read and a future read
204 if (test_and_clear_tsk_thread_flag(task, TIF_PATCH_PENDING))
304 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
390 * Enforce the order of the TIF_PATCH_PENDING read above and the
405 * Sends a fake signal to all non-kthread tasks with TIF_PATCH_PENDING set.
406 * Kthreads with TIF_PATCH_PENDING set are woken up.
421 * There is a small race here. We could see TIF_PATCH_PENDING
489 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
549 set_tsk_thread_flag(task, TIF_PATCH_PENDING);
560 set_tsk_thread_flag(task, TIF_PATCH_PENDING);
587 * has no effect until the TIF_PATCH_PENDING flags get set later.
620 * TIF_PATCH_PENDING writes to ensure klp_update_patch_state() and
659 * Clear all TIF_PATCH_PENDING flags to prevent races caused by
665 clear_tsk_thread_flag(task, TIF_PATCH_PENDING);
669 clear_tsk_thread_flag(idle_task(cpu), TIF_PATCH_PENDING);
673 * __klp_sched_try_switch() see the cleared TIF_PATCH_PENDING before
687 * TIF_PATCH_PENDING writes in klp_start_transition() to ensure
710 if (test_tsk_thread_flag(current, TIF_PATCH_PENDING))
711 set_tsk_thread_flag(child, TIF_PATCH_PENDING);
713 clear_tsk_thread_flag(child, TIF_PATCH_PENDING);
719 * Drop TIF_PATCH_PENDING of all tasks on admin's request. This forces an