Searched refs:thread (Results 1 - 25 of 688) sorted by last modified time

1234567891011>>

/macosx-10.10/libpthread-105.1.4/kern/
H A Dkern_support.c80 #include <kern/thread.h>
115 extern void thread_set_cthreadself(thread_t thread, uint64_t pself, int isLP64);
253 * Pre-fault the first page of the new thread's stack and the page that will
370 /* Set thread QoS class if requested. */
498 /* Unspecified threads means the kernel wants us to impose legacy upon the thread. */
571 /* Static param the thread, we just set QoS on it, so its stuck in QoS land now. */
598 * Drop all outstanding overrides on this thread, done outside the wq lock
632 /* If we have main-thread QoS then we don't allow a thread to come out of QOS_CLASS_UNSPECIFIED. */
698 thread_t thread local
1274 workqueue_callback(int type, thread_t thread) argument
1904 workqueue_run_nextreq(proc_t p, struct workqueue *wq, thread_t thread, boolean_t force_oc, boolean_t overcommit, pthread_priority_t oc_prio) argument
2511 thread_t thread = current_thread(); local
[all...]
/macosx-10.10/libpthread-105.1.4/src/
H A Dpthread.c91 // pthread (other than the main thread) has been created.
118 // Mach message notification that a thread needs to be recycled.
121 pthread_t thread; member in struct:_pthread_reap_msg_t
146 static int _pthread_allocate(pthread_t *thread, const pthread_attr_t *attrs, void **stack);
164 static int _pthread_find_thread(pthread_t thread);
205 extern pthread_t __bsdthread_create(void *(*func)(void *), void * func_arg, void * stack, pthread_t thread, unsigned int flags);
226 // Allocate a thread structure, stack and guard page.
228 // The thread structure may optionally be placed in the same allocation as the
235 // The allocated thread structure is initialized with values that indicate how
239 _pthread_allocate(pthread_t *thread, cons argument
880 pthread_threadid_np(pthread_t thread, uint64_t *thread_id) argument
903 pthread_getname_np(pthread_t thread, char *threadname, size_t len) argument
1059 pthread_create(pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine)(void *), void *arg) argument
1129 pthread_create_suspended_np(pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine)(void *), void *arg) argument
1176 pthread_detach(pthread_t thread) argument
1319 pthread_getschedparam(pthread_t thread, int *policy, struct sched_param *param) argument
1347 pthread_setschedparam_internal(pthread_t thread, mach_port_t kport, int policy, const struct sched_param *param) argument
1483 _pthread_testcancel(pthread_t thread, int isconforming) argument
1599 pthread_t thread = &_thread; local
1709 _pthread_join_cleanup(pthread_t thread, void ** value_ptr, int conforming) argument
1727 _pthread_find_thread(pthread_t thread) argument
1748 _pthread_lookup_thread(pthread_t thread, mach_port_t *portp, int only_joinable) argument
[all...]
H A Dthread_setup.c49 * Machine specific support for thread initialization
57 * Set up the initial state of a MACH thread
60 _pthread_setup(pthread_t thread, argument
83 (void)thread_get_state(_pthread_kernel_thread(thread),
98 *--sp = (uintptr_t)thread; // argument to function
110 state.__rdi = (uintptr_t)thread; // argument to function
123 state.__r[0] = (uintptr_t)thread;
129 (void)thread_set_state(_pthread_kernel_thread(thread), flavor, (thread_state_t)&state, count);
131 (void)thread_resume(_pthread_kernel_thread(thread));
136 _pthread_set_kernel_thread(thread, kernel_threa
[all...]
/macosx-10.10/xnu-2782.1.97/osfmk/vm/
H A Dvm_object.c182 * object concurrently. Only one thread is allowed to
185 * that the first thread has initialized all of the
409 /* I/O Re-prioritization thread */
815 thread_t thread = THREAD_NULL; local
825 result = kernel_thread_start_priority(io_reprioritize_thread, NULL, 95 /* MAXPRI_KERNEL */, &thread);
827 thread_deallocate(thread);
838 thread_t thread; local
844 &thread);
848 thread_deallocate(thread);
1219 "vm_o_deallocate: !cacheable 0x%X res %d paging_ops %d thread
[all...]
H A Dvm_pageout.c89 #include <kern/thread.h>
336 * No locking needed because only one thread modifies the variables.
653 * Wakeup any thread waiting for the page to be un-cleaning.
834 * Given a page, queue it to the appropriate I/O thread,
840 * release the object lock back at the call site. The I/O thread
1097 * thread when it's trying to reclaim memory. We don't need fully
1154 * function in BSD to apply I/O throttle to the pageout thread
1312 * don't want to wake the pageout_scan thread up everytime we fall below
1734 * the pageout thread for the default pager is potentially
1809 * cause us to pause the pageout scan thread
3801 thread_t thread; local
[all...]
H A Dvm_resident.c75 #include <kern/thread.h>
1889 * If another thread allocated space, just bail out now.
1895 * attempt to allocate. Otherwise, a thread will attempt to
1914 * lock to give another thread a chance at it, and
1934 * Return true if it is not likely that a non-vm_privileged thread
2035 * some other thread could have re-filled it... if still
2039 * pageout_scan thread if we moved pages from the global
2119 * must be a privileged thread to be
2121 * thread would have bailed if we were
2297 * We wakeup only one thread, t
[all...]
H A Dvm_compressor.c338 thread_t thread; local
449 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) {
452 thread->options |= TH_OPT_VMPRIV;
454 thread_deallocate(thread);
459 panic("vm_compressor_init: Failed to start the internal pageout thread.\n");
1423 * This function is called from the jetsam thread after killing something to
1682 * thread creation, we want to make sure to delay adjusting
H A Dvm_compressor_backing_store.c173 thread_t thread = NULL; local
189 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) {
192 thread->options |= TH_OPT_VMPRIV;
193 vm_swapout_thread_id = thread->thread_id;
195 thread_deallocate(thread);
198 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) {
201 thread->options |= TH_OPT_VMPRIV;
203 thread_deallocate(thread);
207 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) {
210 thread_deallocate(thread);
[all...]
H A Dvm_fault.c82 #include <kern/thread.h>
126 * will be throttled. The throttling is done by giving the thread that's trying to demand zero a page a
396 * one thread is banging on this object, no problem with the unprotected
554 thread_t thread = current_thread(); local
556 if (thread->options & TH_OPT_VMPRIV)
559 thread->t_page_creation_count++;
565 thread->t_page_creation_count > vm_page_creation_throttle) {
569 elapsed_sec = tv_sec - thread->t_page_creation_time;
571 if (elapsed_sec <= 6 || (thread->t_page_creation_count / elapsed_sec) >= (vm_page_creation_throttle / 6)) {
583 thread
5116 thread_t thread = current_thread(); local
[all...]
H A Dvm_kern.c69 #include <kern/thread.h>
292 * on either end of a stack, they can help detect cases where a thread walks
H A Dvm_map.c4454 * If another thread is wiring/unwiring this entry then
4455 * block after informing other thread to wake us up.
5071 * Another thread is wiring down this entry. Note
5072 * that if it is not for the other thread we would
5078 * Another thread is unwiring this entry. We did not
5707 * Another thread is wiring/unwiring this entry.
5708 * Let the other thread know we are waiting.
5777 * a user thread has physical io pending on
6413 * The kernel-state portion of a user thread must be
8023 thread_t thread local
13228 thread_t thread = current_thread(); local
[all...]
H A Dvm_map.h112 #include <kern/thread.h>
403 * A map copy object may only be used by a single thread
/macosx-10.10/xnu-2782.1.97/osfmk/pmc/
H A Dpmc.c29 #include <kern/thread.h>
218 * Keeps track of all system, task, and thread-level reservations (both active and
221 * We track them all here (rather than in their respective task or thread only)
223 * every task and thread) to determine if/when a new reservation would
703 * or thread to NULL.
743 * Used to place reservation into one of the system, task, and thread queues
772 if(current_thread() == resv->thread) {
800 * This method will inspect the task/thread of the reservation to see if it
801 * matches the new incoming one (for thread/task reservations only). Will only
802 * return TRUE if the task/thread matche
892 pmc_internal_update_thread_flag(thread_t thread, boolean_t newFlag) argument
929 thread_t thread = NULL; local
2469 pmc_reserve_thread(pmc_t pmc, pmc_config_t config, thread_t thread, pmc_reservation_t *reservation) argument
[all...]
H A Dpmc.h402 * reserve the counter using that config in a given execution context (system, or 1 task, or 1 thread),
541 thread_t thread; // not retained member in union:pmc_reservation::__anon15507
664 * @abstract Reserve a PMC for thread-wide counting.
665 * @discussion This method will attempt to reserve the given pmc for thread-wide counting. The resulting reservation will only count when the thread is
669 * @param thread The thread for which to enable the counter.
673 kern_return_t pmc_reserve_thread(pmc_t pmc, pmc_config_t config, thread_t thread, pmc_reservation_t *reservation);
677 * @discussion This method instructs the given reservation to start counting as soon as possible. If the reservation is for a thread (or task) other than the
678 * current thread, o
[all...]
/macosx-10.10/xnu-2782.1.97/osfmk/kern/
H A Dprintf.c161 #include <kern/thread.h>
674 * Lock is only really needed after the first thread is created.
H A Dpriority.c74 #include <kern/thread.h>
86 * Recalculate the quantum and priority for a thread.
97 thread_t thread = p1; local
106 * We bill CPU time to both the individual thread and its task.
109 * thread, we must credit the ledger before taking the thread lock. The ledger
110 * pointers are only manipulated by the thread itself at the ast boundary.
112 ledger_credit(thread->t_ledger, task_ledgers.cpu_time, thread->quantum_remaining);
113 ledger_credit(thread
218 sched_set_thread_base_priority(thread_t thread, int priority) argument
237 lightweight_update_priority(thread_t thread) argument
316 compute_priority( register thread_t thread, boolean_t override_depress) argument
345 compute_my_priority( register thread_t thread) argument
372 can_update_priority( thread_t thread) argument
389 update_priority( register thread_t thread) argument
502 sched_share_incr(thread_t thread) argument
510 sched_share_decr(thread_t thread) argument
519 sched_background_incr(thread_t thread) argument
533 sched_background_decr(thread_t thread) argument
548 assert_thread_sched_count(thread_t thread) argument
588 sched_set_thread_mode(thread_t thread, sched_mode_t new_mode) argument
632 sched_thread_mode_demote(thread_t thread, uint32_t reason) argument
670 sched_thread_mode_undemote(thread_t thread, uint32_t reason) argument
710 sched_set_thread_throttled(thread_t thread, boolean_t wants_throttle) argument
[all...]
H A Dprocessor.c75 #include <kern/thread.h>
441 thread_t thread; local
476 * Create the idle processor thread.
492 * If there is no active thread, the processor
494 * start up thread.
498 result = kernel_thread_create((thread_continue_t)processor_start_thread, NULL, MAXPRI_KERNEL, &thread);
510 thread_lock(thread);
511 thread->bound_processor = processor;
512 processor->next_thread = thread;
513 thread
881 thread_t thread, *thread_list = (thread_t *)addr; local
[all...]
H A Dsched_prim.c100 #include <kern/thread.h>
214 static inline void runq_consider_incr_bound_count(processor_t processor, thread_t thread) argument
216 if (thread->bound_processor == PROCESSOR_NULL)
219 assert(thread->bound_processor == processor);
228 static inline void runq_consider_decr_bound_count(processor_t processor, thread_t thread) argument
230 if (thread->bound_processor == PROCESSOR_NULL)
233 assert(thread->bound_processor == processor);
258 thread_t thread,
264 thread_t thread,
269 thread_t thread,
860 thread_t thread = p0; local
885 thread_unblock( thread_t thread, wait_result_t wresult) argument
1032 thread_go( thread_t thread, wait_result_t wresult) argument
1061 thread_mark_wait_locked( thread_t thread, wait_interrupt_t interruptible) argument
1117 thread_t thread = current_thread(); local
1136 thread_t thread; local
1180 thread_t thread = current_thread(); local
1223 thread_t thread = current_thread(); local
1269 thread_t thread = current_thread(); local
1305 thread_t thread = current_thread(); local
1347 thread_isoncpu(thread_t thread) argument
1384 thread_stop( thread_t thread, boolean_t until_not_runnable) argument
1469 thread_unstop( thread_t thread) argument
1511 thread_wait( thread_t thread, boolean_t until_not_runnable) argument
1576 clear_wait_internal( thread_t thread, wait_result_t wresult) argument
1625 clear_wait( thread_t thread, wait_result_t result) argument
1781 thread_select( thread_t thread, processor_t processor, ast_t reason) argument
1994 thread_select_idle( thread_t thread, processor_t processor) argument
2105 thread_t thread; local
2137 thread_t thread; local
2189 thread_invoke( thread_t self, thread_t thread, ast_t reason) argument
2456 thread_dispatch( thread_t thread, thread_t self) argument
2827 thread_continue( register thread_t thread) argument
2852 thread_quantum_init(thread_t thread) argument
2864 sched_traditional_initial_quantum_size(thread_t thread) argument
2921 sched_traditional_fairshare_enqueue(thread_t thread) argument
2939 thread_t thread; local
2959 sched_traditional_fairshare_queue_remove(thread_t thread) argument
3002 thread_t thread; local
3036 run_queue_enqueue( run_queue_t rq, thread_t thread, integer_t options) argument
3075 run_queue_remove( run_queue_t rq, thread_t thread) argument
3106 fairshare_setrun( processor_t processor, thread_t thread) argument
3130 realtime_queue_insert( thread_t thread) argument
3180 realtime_setrun( processor_t processor, thread_t thread) argument
3308 processor_enqueue( processor_t processor, thread_t thread, integer_t options) argument
3335 processor_setrun( processor_t processor, thread_t thread, integer_t options) argument
3624 choose_processor( processor_set_t pset, processor_t processor, thread_t thread) argument
3927 thread_setrun( thread_t thread, integer_t options) argument
4055 thread_t next, thread; local
4141 thread_t thread = processor->active_thread; local
4187 set_sched_pri( thread_t thread, int priority) argument
4274 processor_queue_remove( processor_t processor, thread_t thread) argument
4329 thread_run_queue_remove( thread_t thread) argument
4405 thread_t thread; local
4455 thread_t thread; local
4511 thread_get_urgency(thread_t thread, uint64_t *arg1, uint64_t *arg2) argument
4560 processor_idle( thread_t thread, processor_t processor) argument
4738 thread_t thread; local
4771 thread_t thread; local
4963 thread_update_add_thread(thread_t thread) argument
4978 thread_t thread = thread_update_array[--thread_update_count]; local
5004 register thread_t thread; local
5036 thread_t thread; local
5075 thread_eager_preemption(thread_t thread) argument
5081 thread_set_eager_preempt(thread_t thread) argument
5115 thread_clear_eager_preempt(thread_t thread) argument
5189 thread_runnable( thread_t thread) argument
[all...]
H A Dsfi.c65 * running a thread that should be SFI-ed are sent an AST.
87 * The wait queue will take the thread's scheduling lock. We may also take
89 * if the thread should block in the wait queue, but the lock will be
348 /* Iterate over processors, call cause_ast_check() on ones running a thread that should be in an off phase */
664 * role of quickly categorizing a thread into its SFI class so that an AST_SFI can be
665 * set. As the thread is unwinding to userspace, sfi_ast() performs full locking
666 * and determines whether the thread should enter an SFI wait state. Because of
668 * thread classification can be inaccurate (but should always be safe). This is
670 * classify the active thread on a remote processor without taking the thread loc
680 sfi_thread_classify(thread_t thread) argument
771 sfi_thread_needs_ast(thread_t thread, sfi_class_id_t *out_class) argument
838 sfi_ast(thread_t thread) argument
924 sfi_reevaluate(thread_t thread) argument
[all...]
H A Dstartup.c92 #include <kern/thread.h>
162 thread_t thread);
240 thread_t thread; local
311 * Initialize the IPC, task, and thread subsystems.
342 * Create a kernel thread to execute the kernel bootstrap.
345 result = kernel_thread_create((thread_continue_t)kernel_bootstrap_thread, NULL, MAXPRI_KERNEL, &thread);
349 thread->state = TH_RUN;
350 thread_deallocate(thread);
353 load_context(thread);
363 * Now running in a thread
593 thread_t thread; local
645 load_context( thread_t thread) argument
[all...]
H A Dtask.c108 #include <kern/thread.h>
277 thread_t thread; local
291 /* FIXME: On x86, the thread save state flavor can diverge from the
294 * certain routines may observe the thread as being in an inconsistent
299 queue_iterate(&task->threads, thread, thread_t, task_threads) {
300 thread_mtx_lock(thread);
301 machine_thread_switch_addrmode(thread);
302 thread_mtx_unlock(thread);
1165 * (kern/thread.c) about problems with terminating the "current task."
1209 thread_t thread, sel local
1400 thread_t thread, self; local
1519 register thread_t thread; local
1603 register thread_t thread, self; local
1632 register thread_t thread; local
1684 thread_t thread; local
2642 register thread_t thread; local
2675 register thread_t thread; local
2953 register thread_t thread; local
3172 thread_t thread; local
3235 thread_t thread; local
3273 thread_t thread; local
3344 thread_t thread = current_thread(); local
3833 thread_t thread= THREAD_NULL; local
[all...]
H A Dtask.h149 #include <kern/thread.h>
265 #define TF_CPUMON_WARNING 0x00000004 /* task has at least one thread in CPU usage warning zone */
299 uint8_t rusage_cpu_perthr_percentage; /* Per-thread CPU limit percentage */
300 uint64_t rusage_cpu_perthr_interval; /* Per-thread CPU limit interval */
556 extern kern_return_t check_actforsig(task_t task, thread_t thread, int setast);
614 /* internal or external, thread or task */
641 /* thread-only attributes */
651 extern void proc_set_task_policy(task_t task, thread_t thread, int category, int flavor, int value);
652 extern int proc_get_task_policy(task_t task, thread_t thread, int category, int flavor);
655 extern void proc_set_task_policy2(task_t task, thread_t thread, in
[all...]
H A Dtask_policy.c57 * This subsystem manages task and thread IO priority and backgrounding,
77 * Here are the steps to change a policy on a task or thread:
108 * Changing task policy on a task or thread takes the task lock, and not the thread lock.
109 * TODO: Should changing policy on a thread take the thread lock instead?
123 extern void thread_recompute_qos(thread_t thread);
126 static void proc_set_task_policy_locked(task_t task, thread_t thread, int category, int flavor, int value);
127 static void proc_set_task_policy2_locked(task_t task, thread_t thread, int category, int flavor, int value1, int value2);
129 static void task_policy_update_locked(task_t task, thread_t thread, task_pend_token_
670 thread_policy_create(thread_t thread) argument
688 task_policy_update_locked(task_t task, thread_t thread, task_pend_token_t pend_token) argument
716 task_policy_update_internal_locked(task_t task, thread_t thread, boolean_t in_create, task_pend_token_t pend_token) argument
1202 task_policy_update_thread_locked(thread_t thread, int update_cpu, boolean_t update_throttle, boolean_t update_sfi, boolean_t update_qos) argument
1307 thread_t thread; local
1341 task_policy_update_complete_unlocked(task_t task, thread_t thread, task_pend_token_t pend_token) argument
1372 proc_set_task_policy(task_t task, thread_t thread, int category, int flavor, int value) argument
1410 thread_t thread; local
1448 proc_set_task_policy2(task_t task, thread_t thread, int category, int flavor, int value1, int value2) argument
1478 proc_set_task_policy_locked(task_t task, thread_t thread, int category, int flavor, int value) argument
1606 proc_set_task_policy2_locked(task_t task, thread_t thread, int category, int flavor, int value1, int value2) argument
1660 proc_get_task_policy(task_t task, thread_t thread, int category, int flavor) argument
1735 proc_get_task_policy2(task_t task, thread_t thread, int category __unused, int flavor, int *value1, int *value2) argument
1787 proc_get_effective_thread_policy(thread_t thread, int flavor) argument
1801 proc_get_effective_policy(task_t task, thread_t thread, int flavor) argument
2016 proc_apply_workq_bgthreadpolicy(thread_t thread) argument
2032 proc_restore_workq_bgthreadpolicy(thread_t thread) argument
2045 proc_setthread_saved_importance(__unused thread_t thread, __unused int importance) argument
2065 set_thread_iotier_override(thread_t thread, int policy) argument
2104 proc_thread_qos_add_override(task_t task, thread_t thread, uint64_t tid, int override_qos, boolean_t first_override_for_resource) argument
2175 proc_thread_qos_remove_override(task_t task, thread_t thread, uint64_t tid) argument
2457 proc_get_thread_policy(thread_t thread, thread_policy_state_t info) argument
2510 trequested_0(task_t task, thread_t thread) argument
2521 trequested_1(task_t task, thread_t thread) argument
2536 teffective_0(task_t task, thread_t thread) argument
2547 teffective_1(task_t task, thread_t thread) argument
2565 task_requested_bitfield(task_t task, thread_t thread) argument
2613 task_effective_bitfield(task_t task, thread_t thread) argument
2901 thread_t thread; local
2933 thread_t thread; local
[all...]
H A Dtelemetry.c76 void telemetry_take_sample(thread_t thread, uint8_t microsnapshot_flags, struct micro_snapshot_buffer * current_buffer);
331 * Determine if the current thread is eligible for telemetry:
338 telemetry_is_active(thread_t thread) argument
344 if ((telemetry_active_tasks > 0) && ((thread->task->t_flags & TF_TELEMETRY) != 0)) {
367 * Mark the current thread for an interrupt-based
373 thread_t thread = current_thread(); local
376 * If telemetry isn't active for this thread, return and try
379 if (telemetry_is_active(thread) == FALSE) {
391 thread_ast_set(thread, ast_bits);
392 ast_propagate(thread
449 telemetry_ast(thread_t thread, boolean_t interrupted_userspace, boolean_t is_windowed) argument
463 telemetry_take_sample(thread_t thread, uint8_t microsnapshot_flags, struct micro_snapshot_buffer * current_buffer) argument
[all...]
H A Dthread.c59 * File: kern/thread.c
93 #include <machine/thread.h>
114 #include <kern/thread.h>
155 static struct thread thread_template, init_thread;
159 thread_t thread);
188 * (ie when any thread's CPU consumption exceeds 70% of the limit, start taking user
205 * Fill in a template thread for fast initialization.
353 sizeof(struct thread),
354 thread_max * sizeof(struct thread),
355 THREAD_CHUNK * sizeof(struct thread),
393 thread_t thread = current_thread(); local
510 thread_deallocate( thread_t thread) argument
571 thread_t self, thread; local
649 thread_terminate_enqueue( thread_t thread) argument
670 thread_t thread; local
707 thread_stack_enqueue( thread_t thread) argument
721 thread_t thread = NULL; local
977 thread_t thread; local
1034 thread_t thread; local
1111 thread_t thread; local
1147 thread_t thread; local
1178 thread_t thread; local
1204 thread_info_internal( register thread_t thread, thread_flavor_t flavor, thread_info_t thread_info_out, mach_msg_type_number_t *thread_info_count) argument
1416 thread_read_times( thread_t thread, time_value_t *user_time, time_value_t *system_time) argument
1452 thread_t thread = NULL; local
1468 thread_assign( __unused thread_t thread, __unused processor_set_t new_pset) argument
1482 thread_assign_default( thread_t thread) argument
1494 thread_get_assignment( thread_t thread, processor_set_t *pset) argument
1513 thread_wire_internal( host_priv_t host_priv, thread_t thread, boolean_t wired, boolean_t *prev_state) argument
1548 thread_wire( host_priv_t host_priv, thread_t thread, boolean_t wired) argument
1580 thread_guard_violation(thread_t thread, unsigned type) argument
1607 guard_ast(thread_t thread) argument
1650 thread_t thread = current_thread(); local
1776 thread_update_io_stats(thread_t thread, int size, int io_flags) argument
1841 thread_t thread = current_thread(); local
1896 thread_t thread = current_thread(); local
1981 sched_call_null( __unused int type, __unused thread_t thread) argument
1989 thread_sched_call( thread_t thread, sched_call_t call) argument
1997 thread_static_param( thread_t thread, boolean_t state) argument
2007 thread_tid( thread_t thread) argument
2021 thread_dispatchqaddr( thread_t thread) argument
2045 thread_reference( thread_t thread) argument
2074 thread_t thread = current_thread(); local
2131 thread_get_mach_voucher( thread_act_t thread, mach_voucher_selector_t __unused which, ipc_voucher_t *voucherp) argument
2201 thread_set_mach_voucher( thread_t thread, ipc_voucher_t voucher) argument
2253 thread_swap_mach_voucher( thread_t thread, ipc_voucher_t new_voucher, ipc_voucher_t *in_out_old_voucher) argument
2336 thread_t thread = current_thread(); local
2351 dtrace_get_thread_predcache(thread_t thread) argument
2359 dtrace_get_thread_vtime(thread_t thread) argument
2367 dtrace_get_thread_tracing(thread_t thread) argument
2375 dtrace_get_thread_reentering(thread_t thread) argument
2383 dtrace_get_kernel_stack(thread_t thread) argument
2391 dtrace_calc_thread_recent_vtime(thread_t thread) argument
2406 dtrace_set_thread_predcache(thread_t thread, uint32_t predcache) argument
2412 dtrace_set_thread_vtime(thread_t thread, int64_t vtime) argument
2418 dtrace_set_thread_tracing(thread_t thread, int64_t accum) argument
2424 dtrace_set_thread_reentering(thread_t thread, boolean_t vbool) argument
2434 dtrace_set_thread_recover(thread_t thread, vm_offset_t recover) argument
2450 thread_t thread = current_thread(); local
2462 dtrace_thread_didexec(thread_t thread) argument
[all...]

Completed in 333 milliseconds

1234567891011>>