/macosx-10.10/libpthread-105.1.4/kern/ |
H A D | kern_support.c | 80 #include <kern/thread.h> 115 extern void thread_set_cthreadself(thread_t thread, uint64_t pself, int isLP64); 253 * Pre-fault the first page of the new thread's stack and the page that will 370 /* Set thread QoS class if requested. */ 498 /* Unspecified threads means the kernel wants us to impose legacy upon the thread. */ 571 /* Static param the thread, we just set QoS on it, so its stuck in QoS land now. */ 598 * Drop all outstanding overrides on this thread, done outside the wq lock 632 /* If we have main-thread QoS then we don't allow a thread to come out of QOS_CLASS_UNSPECIFIED. */ 698 thread_t thread local 1274 workqueue_callback(int type, thread_t thread) argument 1904 workqueue_run_nextreq(proc_t p, struct workqueue *wq, thread_t thread, boolean_t force_oc, boolean_t overcommit, pthread_priority_t oc_prio) argument 2511 thread_t thread = current_thread(); local [all...] |
/macosx-10.10/libpthread-105.1.4/src/ |
H A D | pthread.c | 91 // pthread (other than the main thread) has been created. 118 // Mach message notification that a thread needs to be recycled. 121 pthread_t thread; member in struct:_pthread_reap_msg_t 146 static int _pthread_allocate(pthread_t *thread, const pthread_attr_t *attrs, void **stack); 164 static int _pthread_find_thread(pthread_t thread); 205 extern pthread_t __bsdthread_create(void *(*func)(void *), void * func_arg, void * stack, pthread_t thread, unsigned int flags); 226 // Allocate a thread structure, stack and guard page. 228 // The thread structure may optionally be placed in the same allocation as the 235 // The allocated thread structure is initialized with values that indicate how 239 _pthread_allocate(pthread_t *thread, cons argument 880 pthread_threadid_np(pthread_t thread, uint64_t *thread_id) argument 903 pthread_getname_np(pthread_t thread, char *threadname, size_t len) argument 1059 pthread_create(pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine)(void *), void *arg) argument 1129 pthread_create_suspended_np(pthread_t *thread, const pthread_attr_t *attr, void *(*start_routine)(void *), void *arg) argument 1176 pthread_detach(pthread_t thread) argument 1319 pthread_getschedparam(pthread_t thread, int *policy, struct sched_param *param) argument 1347 pthread_setschedparam_internal(pthread_t thread, mach_port_t kport, int policy, const struct sched_param *param) argument 1483 _pthread_testcancel(pthread_t thread, int isconforming) argument 1599 pthread_t thread = &_thread; local 1709 _pthread_join_cleanup(pthread_t thread, void ** value_ptr, int conforming) argument 1727 _pthread_find_thread(pthread_t thread) argument 1748 _pthread_lookup_thread(pthread_t thread, mach_port_t *portp, int only_joinable) argument [all...] |
H A D | thread_setup.c | 49 * Machine specific support for thread initialization 57 * Set up the initial state of a MACH thread 60 _pthread_setup(pthread_t thread, argument 83 (void)thread_get_state(_pthread_kernel_thread(thread), 98 *--sp = (uintptr_t)thread; // argument to function 110 state.__rdi = (uintptr_t)thread; // argument to function 123 state.__r[0] = (uintptr_t)thread; 129 (void)thread_set_state(_pthread_kernel_thread(thread), flavor, (thread_state_t)&state, count); 131 (void)thread_resume(_pthread_kernel_thread(thread)); 136 _pthread_set_kernel_thread(thread, kernel_threa [all...] |
/macosx-10.10/xnu-2782.1.97/osfmk/vm/ |
H A D | vm_object.c | 182 * object concurrently. Only one thread is allowed to 185 * that the first thread has initialized all of the 409 /* I/O Re-prioritization thread */ 815 thread_t thread = THREAD_NULL; local 825 result = kernel_thread_start_priority(io_reprioritize_thread, NULL, 95 /* MAXPRI_KERNEL */, &thread); 827 thread_deallocate(thread); 838 thread_t thread; local 844 &thread); 848 thread_deallocate(thread); 1219 "vm_o_deallocate: !cacheable 0x%X res %d paging_ops %d thread [all...] |
H A D | vm_pageout.c | 89 #include <kern/thread.h> 336 * No locking needed because only one thread modifies the variables. 653 * Wakeup any thread waiting for the page to be un-cleaning. 834 * Given a page, queue it to the appropriate I/O thread, 840 * release the object lock back at the call site. The I/O thread 1097 * thread when it's trying to reclaim memory. We don't need fully 1154 * function in BSD to apply I/O throttle to the pageout thread 1312 * don't want to wake the pageout_scan thread up everytime we fall below 1734 * the pageout thread for the default pager is potentially 1809 * cause us to pause the pageout scan thread 3801 thread_t thread; local [all...] |
H A D | vm_resident.c | 75 #include <kern/thread.h> 1889 * If another thread allocated space, just bail out now. 1895 * attempt to allocate. Otherwise, a thread will attempt to 1914 * lock to give another thread a chance at it, and 1934 * Return true if it is not likely that a non-vm_privileged thread 2035 * some other thread could have re-filled it... if still 2039 * pageout_scan thread if we moved pages from the global 2119 * must be a privileged thread to be 2121 * thread would have bailed if we were 2297 * We wakeup only one thread, t [all...] |
H A D | vm_compressor.c | 338 thread_t thread; local 449 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) { 452 thread->options |= TH_OPT_VMPRIV; 454 thread_deallocate(thread); 459 panic("vm_compressor_init: Failed to start the internal pageout thread.\n"); 1423 * This function is called from the jetsam thread after killing something to 1682 * thread creation, we want to make sure to delay adjusting
|
H A D | vm_compressor_backing_store.c | 173 thread_t thread = NULL; local 189 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) { 192 thread->options |= TH_OPT_VMPRIV; 193 vm_swapout_thread_id = thread->thread_id; 195 thread_deallocate(thread); 198 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) { 201 thread->options |= TH_OPT_VMPRIV; 203 thread_deallocate(thread); 207 BASEPRI_PREEMPT - 1, &thread) != KERN_SUCCESS) { 210 thread_deallocate(thread); [all...] |
H A D | vm_fault.c | 82 #include <kern/thread.h> 126 * will be throttled. The throttling is done by giving the thread that's trying to demand zero a page a 396 * one thread is banging on this object, no problem with the unprotected 554 thread_t thread = current_thread(); local 556 if (thread->options & TH_OPT_VMPRIV) 559 thread->t_page_creation_count++; 565 thread->t_page_creation_count > vm_page_creation_throttle) { 569 elapsed_sec = tv_sec - thread->t_page_creation_time; 571 if (elapsed_sec <= 6 || (thread->t_page_creation_count / elapsed_sec) >= (vm_page_creation_throttle / 6)) { 583 thread 5116 thread_t thread = current_thread(); local [all...] |
H A D | vm_kern.c | 69 #include <kern/thread.h> 292 * on either end of a stack, they can help detect cases where a thread walks
|
H A D | vm_map.c | 4454 * If another thread is wiring/unwiring this entry then 4455 * block after informing other thread to wake us up. 5071 * Another thread is wiring down this entry. Note 5072 * that if it is not for the other thread we would 5078 * Another thread is unwiring this entry. We did not 5707 * Another thread is wiring/unwiring this entry. 5708 * Let the other thread know we are waiting. 5777 * a user thread has physical io pending on 6413 * The kernel-state portion of a user thread must be 8023 thread_t thread local 13228 thread_t thread = current_thread(); local [all...] |
H A D | vm_map.h | 112 #include <kern/thread.h> 403 * A map copy object may only be used by a single thread
|
/macosx-10.10/xnu-2782.1.97/osfmk/pmc/ |
H A D | pmc.c | 29 #include <kern/thread.h> 218 * Keeps track of all system, task, and thread-level reservations (both active and 221 * We track them all here (rather than in their respective task or thread only) 223 * every task and thread) to determine if/when a new reservation would 703 * or thread to NULL. 743 * Used to place reservation into one of the system, task, and thread queues 772 if(current_thread() == resv->thread) { 800 * This method will inspect the task/thread of the reservation to see if it 801 * matches the new incoming one (for thread/task reservations only). Will only 802 * return TRUE if the task/thread matche 892 pmc_internal_update_thread_flag(thread_t thread, boolean_t newFlag) argument 929 thread_t thread = NULL; local 2469 pmc_reserve_thread(pmc_t pmc, pmc_config_t config, thread_t thread, pmc_reservation_t *reservation) argument [all...] |
H A D | pmc.h | 402 * reserve the counter using that config in a given execution context (system, or 1 task, or 1 thread), 541 thread_t thread; // not retained member in union:pmc_reservation::__anon15507 664 * @abstract Reserve a PMC for thread-wide counting. 665 * @discussion This method will attempt to reserve the given pmc for thread-wide counting. The resulting reservation will only count when the thread is 669 * @param thread The thread for which to enable the counter. 673 kern_return_t pmc_reserve_thread(pmc_t pmc, pmc_config_t config, thread_t thread, pmc_reservation_t *reservation); 677 * @discussion This method instructs the given reservation to start counting as soon as possible. If the reservation is for a thread (or task) other than the 678 * current thread, o [all...] |
/macosx-10.10/xnu-2782.1.97/osfmk/kern/ |
H A D | printf.c | 161 #include <kern/thread.h> 674 * Lock is only really needed after the first thread is created.
|
H A D | priority.c | 74 #include <kern/thread.h> 86 * Recalculate the quantum and priority for a thread. 97 thread_t thread = p1; local 106 * We bill CPU time to both the individual thread and its task. 109 * thread, we must credit the ledger before taking the thread lock. The ledger 110 * pointers are only manipulated by the thread itself at the ast boundary. 112 ledger_credit(thread->t_ledger, task_ledgers.cpu_time, thread->quantum_remaining); 113 ledger_credit(thread 218 sched_set_thread_base_priority(thread_t thread, int priority) argument 237 lightweight_update_priority(thread_t thread) argument 316 compute_priority( register thread_t thread, boolean_t override_depress) argument 345 compute_my_priority( register thread_t thread) argument 372 can_update_priority( thread_t thread) argument 389 update_priority( register thread_t thread) argument 502 sched_share_incr(thread_t thread) argument 510 sched_share_decr(thread_t thread) argument 519 sched_background_incr(thread_t thread) argument 533 sched_background_decr(thread_t thread) argument 548 assert_thread_sched_count(thread_t thread) argument 588 sched_set_thread_mode(thread_t thread, sched_mode_t new_mode) argument 632 sched_thread_mode_demote(thread_t thread, uint32_t reason) argument 670 sched_thread_mode_undemote(thread_t thread, uint32_t reason) argument 710 sched_set_thread_throttled(thread_t thread, boolean_t wants_throttle) argument [all...] |
H A D | processor.c | 75 #include <kern/thread.h> 441 thread_t thread; local 476 * Create the idle processor thread. 492 * If there is no active thread, the processor 494 * start up thread. 498 result = kernel_thread_create((thread_continue_t)processor_start_thread, NULL, MAXPRI_KERNEL, &thread); 510 thread_lock(thread); 511 thread->bound_processor = processor; 512 processor->next_thread = thread; 513 thread 881 thread_t thread, *thread_list = (thread_t *)addr; local [all...] |
H A D | sched_prim.c | 100 #include <kern/thread.h> 214 static inline void runq_consider_incr_bound_count(processor_t processor, thread_t thread) argument 216 if (thread->bound_processor == PROCESSOR_NULL) 219 assert(thread->bound_processor == processor); 228 static inline void runq_consider_decr_bound_count(processor_t processor, thread_t thread) argument 230 if (thread->bound_processor == PROCESSOR_NULL) 233 assert(thread->bound_processor == processor); 258 thread_t thread, 264 thread_t thread, 269 thread_t thread, 860 thread_t thread = p0; local 885 thread_unblock( thread_t thread, wait_result_t wresult) argument 1032 thread_go( thread_t thread, wait_result_t wresult) argument 1061 thread_mark_wait_locked( thread_t thread, wait_interrupt_t interruptible) argument 1117 thread_t thread = current_thread(); local 1136 thread_t thread; local 1180 thread_t thread = current_thread(); local 1223 thread_t thread = current_thread(); local 1269 thread_t thread = current_thread(); local 1305 thread_t thread = current_thread(); local 1347 thread_isoncpu(thread_t thread) argument 1384 thread_stop( thread_t thread, boolean_t until_not_runnable) argument 1469 thread_unstop( thread_t thread) argument 1511 thread_wait( thread_t thread, boolean_t until_not_runnable) argument 1576 clear_wait_internal( thread_t thread, wait_result_t wresult) argument 1625 clear_wait( thread_t thread, wait_result_t result) argument 1781 thread_select( thread_t thread, processor_t processor, ast_t reason) argument 1994 thread_select_idle( thread_t thread, processor_t processor) argument 2105 thread_t thread; local 2137 thread_t thread; local 2189 thread_invoke( thread_t self, thread_t thread, ast_t reason) argument 2456 thread_dispatch( thread_t thread, thread_t self) argument 2827 thread_continue( register thread_t thread) argument 2852 thread_quantum_init(thread_t thread) argument 2864 sched_traditional_initial_quantum_size(thread_t thread) argument 2921 sched_traditional_fairshare_enqueue(thread_t thread) argument 2939 thread_t thread; local 2959 sched_traditional_fairshare_queue_remove(thread_t thread) argument 3002 thread_t thread; local 3036 run_queue_enqueue( run_queue_t rq, thread_t thread, integer_t options) argument 3075 run_queue_remove( run_queue_t rq, thread_t thread) argument 3106 fairshare_setrun( processor_t processor, thread_t thread) argument 3130 realtime_queue_insert( thread_t thread) argument 3180 realtime_setrun( processor_t processor, thread_t thread) argument 3308 processor_enqueue( processor_t processor, thread_t thread, integer_t options) argument 3335 processor_setrun( processor_t processor, thread_t thread, integer_t options) argument 3624 choose_processor( processor_set_t pset, processor_t processor, thread_t thread) argument 3927 thread_setrun( thread_t thread, integer_t options) argument 4055 thread_t next, thread; local 4141 thread_t thread = processor->active_thread; local 4187 set_sched_pri( thread_t thread, int priority) argument 4274 processor_queue_remove( processor_t processor, thread_t thread) argument 4329 thread_run_queue_remove( thread_t thread) argument 4405 thread_t thread; local 4455 thread_t thread; local 4511 thread_get_urgency(thread_t thread, uint64_t *arg1, uint64_t *arg2) argument 4560 processor_idle( thread_t thread, processor_t processor) argument 4738 thread_t thread; local 4771 thread_t thread; local 4963 thread_update_add_thread(thread_t thread) argument 4978 thread_t thread = thread_update_array[--thread_update_count]; local 5004 register thread_t thread; local 5036 thread_t thread; local 5075 thread_eager_preemption(thread_t thread) argument 5081 thread_set_eager_preempt(thread_t thread) argument 5115 thread_clear_eager_preempt(thread_t thread) argument 5189 thread_runnable( thread_t thread) argument [all...] |
H A D | sfi.c | 65 * running a thread that should be SFI-ed are sent an AST. 87 * The wait queue will take the thread's scheduling lock. We may also take 89 * if the thread should block in the wait queue, but the lock will be 348 /* Iterate over processors, call cause_ast_check() on ones running a thread that should be in an off phase */ 664 * role of quickly categorizing a thread into its SFI class so that an AST_SFI can be 665 * set. As the thread is unwinding to userspace, sfi_ast() performs full locking 666 * and determines whether the thread should enter an SFI wait state. Because of 668 * thread classification can be inaccurate (but should always be safe). This is 670 * classify the active thread on a remote processor without taking the thread loc 680 sfi_thread_classify(thread_t thread) argument 771 sfi_thread_needs_ast(thread_t thread, sfi_class_id_t *out_class) argument 838 sfi_ast(thread_t thread) argument 924 sfi_reevaluate(thread_t thread) argument [all...] |
H A D | startup.c | 92 #include <kern/thread.h> 162 thread_t thread); 240 thread_t thread; local 311 * Initialize the IPC, task, and thread subsystems. 342 * Create a kernel thread to execute the kernel bootstrap. 345 result = kernel_thread_create((thread_continue_t)kernel_bootstrap_thread, NULL, MAXPRI_KERNEL, &thread); 349 thread->state = TH_RUN; 350 thread_deallocate(thread); 353 load_context(thread); 363 * Now running in a thread 593 thread_t thread; local 645 load_context( thread_t thread) argument [all...] |
H A D | task.c | 108 #include <kern/thread.h> 277 thread_t thread; local 291 /* FIXME: On x86, the thread save state flavor can diverge from the 294 * certain routines may observe the thread as being in an inconsistent 299 queue_iterate(&task->threads, thread, thread_t, task_threads) { 300 thread_mtx_lock(thread); 301 machine_thread_switch_addrmode(thread); 302 thread_mtx_unlock(thread); 1165 * (kern/thread.c) about problems with terminating the "current task." 1209 thread_t thread, sel local 1400 thread_t thread, self; local 1519 register thread_t thread; local 1603 register thread_t thread, self; local 1632 register thread_t thread; local 1684 thread_t thread; local 2642 register thread_t thread; local 2675 register thread_t thread; local 2953 register thread_t thread; local 3172 thread_t thread; local 3235 thread_t thread; local 3273 thread_t thread; local 3344 thread_t thread = current_thread(); local 3833 thread_t thread= THREAD_NULL; local [all...] |
H A D | task.h | 149 #include <kern/thread.h> 265 #define TF_CPUMON_WARNING 0x00000004 /* task has at least one thread in CPU usage warning zone */ 299 uint8_t rusage_cpu_perthr_percentage; /* Per-thread CPU limit percentage */ 300 uint64_t rusage_cpu_perthr_interval; /* Per-thread CPU limit interval */ 556 extern kern_return_t check_actforsig(task_t task, thread_t thread, int setast); 614 /* internal or external, thread or task */ 641 /* thread-only attributes */ 651 extern void proc_set_task_policy(task_t task, thread_t thread, int category, int flavor, int value); 652 extern int proc_get_task_policy(task_t task, thread_t thread, int category, int flavor); 655 extern void proc_set_task_policy2(task_t task, thread_t thread, in [all...] |
H A D | task_policy.c | 57 * This subsystem manages task and thread IO priority and backgrounding, 77 * Here are the steps to change a policy on a task or thread: 108 * Changing task policy on a task or thread takes the task lock, and not the thread lock. 109 * TODO: Should changing policy on a thread take the thread lock instead? 123 extern void thread_recompute_qos(thread_t thread); 126 static void proc_set_task_policy_locked(task_t task, thread_t thread, int category, int flavor, int value); 127 static void proc_set_task_policy2_locked(task_t task, thread_t thread, int category, int flavor, int value1, int value2); 129 static void task_policy_update_locked(task_t task, thread_t thread, task_pend_token_ 670 thread_policy_create(thread_t thread) argument 688 task_policy_update_locked(task_t task, thread_t thread, task_pend_token_t pend_token) argument 716 task_policy_update_internal_locked(task_t task, thread_t thread, boolean_t in_create, task_pend_token_t pend_token) argument 1202 task_policy_update_thread_locked(thread_t thread, int update_cpu, boolean_t update_throttle, boolean_t update_sfi, boolean_t update_qos) argument 1307 thread_t thread; local 1341 task_policy_update_complete_unlocked(task_t task, thread_t thread, task_pend_token_t pend_token) argument 1372 proc_set_task_policy(task_t task, thread_t thread, int category, int flavor, int value) argument 1410 thread_t thread; local 1448 proc_set_task_policy2(task_t task, thread_t thread, int category, int flavor, int value1, int value2) argument 1478 proc_set_task_policy_locked(task_t task, thread_t thread, int category, int flavor, int value) argument 1606 proc_set_task_policy2_locked(task_t task, thread_t thread, int category, int flavor, int value1, int value2) argument 1660 proc_get_task_policy(task_t task, thread_t thread, int category, int flavor) argument 1735 proc_get_task_policy2(task_t task, thread_t thread, int category __unused, int flavor, int *value1, int *value2) argument 1787 proc_get_effective_thread_policy(thread_t thread, int flavor) argument 1801 proc_get_effective_policy(task_t task, thread_t thread, int flavor) argument 2016 proc_apply_workq_bgthreadpolicy(thread_t thread) argument 2032 proc_restore_workq_bgthreadpolicy(thread_t thread) argument 2045 proc_setthread_saved_importance(__unused thread_t thread, __unused int importance) argument 2065 set_thread_iotier_override(thread_t thread, int policy) argument 2104 proc_thread_qos_add_override(task_t task, thread_t thread, uint64_t tid, int override_qos, boolean_t first_override_for_resource) argument 2175 proc_thread_qos_remove_override(task_t task, thread_t thread, uint64_t tid) argument 2457 proc_get_thread_policy(thread_t thread, thread_policy_state_t info) argument 2510 trequested_0(task_t task, thread_t thread) argument 2521 trequested_1(task_t task, thread_t thread) argument 2536 teffective_0(task_t task, thread_t thread) argument 2547 teffective_1(task_t task, thread_t thread) argument 2565 task_requested_bitfield(task_t task, thread_t thread) argument 2613 task_effective_bitfield(task_t task, thread_t thread) argument 2901 thread_t thread; local 2933 thread_t thread; local [all...] |
H A D | telemetry.c | 76 void telemetry_take_sample(thread_t thread, uint8_t microsnapshot_flags, struct micro_snapshot_buffer * current_buffer); 331 * Determine if the current thread is eligible for telemetry: 338 telemetry_is_active(thread_t thread) argument 344 if ((telemetry_active_tasks > 0) && ((thread->task->t_flags & TF_TELEMETRY) != 0)) { 367 * Mark the current thread for an interrupt-based 373 thread_t thread = current_thread(); local 376 * If telemetry isn't active for this thread, return and try 379 if (telemetry_is_active(thread) == FALSE) { 391 thread_ast_set(thread, ast_bits); 392 ast_propagate(thread 449 telemetry_ast(thread_t thread, boolean_t interrupted_userspace, boolean_t is_windowed) argument 463 telemetry_take_sample(thread_t thread, uint8_t microsnapshot_flags, struct micro_snapshot_buffer * current_buffer) argument [all...] |
H A D | thread.c | 59 * File: kern/thread.c 93 #include <machine/thread.h> 114 #include <kern/thread.h> 155 static struct thread thread_template, init_thread; 159 thread_t thread); 188 * (ie when any thread's CPU consumption exceeds 70% of the limit, start taking user 205 * Fill in a template thread for fast initialization. 353 sizeof(struct thread), 354 thread_max * sizeof(struct thread), 355 THREAD_CHUNK * sizeof(struct thread), 393 thread_t thread = current_thread(); local 510 thread_deallocate( thread_t thread) argument 571 thread_t self, thread; local 649 thread_terminate_enqueue( thread_t thread) argument 670 thread_t thread; local 707 thread_stack_enqueue( thread_t thread) argument 721 thread_t thread = NULL; local 977 thread_t thread; local 1034 thread_t thread; local 1111 thread_t thread; local 1147 thread_t thread; local 1178 thread_t thread; local 1204 thread_info_internal( register thread_t thread, thread_flavor_t flavor, thread_info_t thread_info_out, mach_msg_type_number_t *thread_info_count) argument 1416 thread_read_times( thread_t thread, time_value_t *user_time, time_value_t *system_time) argument 1452 thread_t thread = NULL; local 1468 thread_assign( __unused thread_t thread, __unused processor_set_t new_pset) argument 1482 thread_assign_default( thread_t thread) argument 1494 thread_get_assignment( thread_t thread, processor_set_t *pset) argument 1513 thread_wire_internal( host_priv_t host_priv, thread_t thread, boolean_t wired, boolean_t *prev_state) argument 1548 thread_wire( host_priv_t host_priv, thread_t thread, boolean_t wired) argument 1580 thread_guard_violation(thread_t thread, unsigned type) argument 1607 guard_ast(thread_t thread) argument 1650 thread_t thread = current_thread(); local 1776 thread_update_io_stats(thread_t thread, int size, int io_flags) argument 1841 thread_t thread = current_thread(); local 1896 thread_t thread = current_thread(); local 1981 sched_call_null( __unused int type, __unused thread_t thread) argument 1989 thread_sched_call( thread_t thread, sched_call_t call) argument 1997 thread_static_param( thread_t thread, boolean_t state) argument 2007 thread_tid( thread_t thread) argument 2021 thread_dispatchqaddr( thread_t thread) argument 2045 thread_reference( thread_t thread) argument 2074 thread_t thread = current_thread(); local 2131 thread_get_mach_voucher( thread_act_t thread, mach_voucher_selector_t __unused which, ipc_voucher_t *voucherp) argument 2201 thread_set_mach_voucher( thread_t thread, ipc_voucher_t voucher) argument 2253 thread_swap_mach_voucher( thread_t thread, ipc_voucher_t new_voucher, ipc_voucher_t *in_out_old_voucher) argument 2336 thread_t thread = current_thread(); local 2351 dtrace_get_thread_predcache(thread_t thread) argument 2359 dtrace_get_thread_vtime(thread_t thread) argument 2367 dtrace_get_thread_tracing(thread_t thread) argument 2375 dtrace_get_thread_reentering(thread_t thread) argument 2383 dtrace_get_kernel_stack(thread_t thread) argument 2391 dtrace_calc_thread_recent_vtime(thread_t thread) argument 2406 dtrace_set_thread_predcache(thread_t thread, uint32_t predcache) argument 2412 dtrace_set_thread_vtime(thread_t thread, int64_t vtime) argument 2418 dtrace_set_thread_tracing(thread_t thread, int64_t accum) argument 2424 dtrace_set_thread_reentering(thread_t thread, boolean_t vbool) argument 2434 dtrace_set_thread_recover(thread_t thread, vm_offset_t recover) argument 2450 thread_t thread = current_thread(); local 2462 dtrace_thread_didexec(thread_t thread) argument [all...] |