Lines Matching refs:td

131 	struct thread *td;
138 td = curthread;
139 p = td->td_proc;
141 if (KTRPOINT(td, KTR_CSW))
149 KASSERT(ident != NULL && TD_IS_RUNNING(td), ("msleep"));
158 if (SCHEDULER_STOPPED_TD(td)) {
166 KASSERT(!TD_ON_SLEEPQ(td), ("recursive sleep"));
178 td->td_tid, p->p_pid, td->td_name, wmesg, ident);
197 * stopped, then td will no longer be on a sleep queue upon
220 if (KTRPOINT(td, KTR_CSW))
235 struct thread *td;
240 td = curthread;
241 p = td->td_proc;
244 KASSERT(ident != NULL && TD_IS_RUNNING(td), ("msleep"));
246 if (SCHEDULER_STOPPED_TD(td))
251 td->td_tid, p->p_pid, td->td_name, wmesg, ident);
273 if (KTRPOINT(td, KTR_CSW)) {
292 if (KTRPOINT(td, KTR_CSW))
401 struct thread *td;
403 td = curthread; /* XXX */
404 THREAD_LOCK_ASSERT(td, MA_OWNED | MA_NOTRECURSED);
405 KASSERT(!TD_ON_RUNQ(td), ("mi_switch: called by old code"));
407 if (!TD_ON_LOCK(td) && !TD_IS_RUNNING(td))
410 KASSERT(td->td_critnest == 1 || panicstr,
421 if (SCHEDULER_STOPPED_TD(td))
424 td->td_ru.ru_nvcsw++;
425 td->td_swvoltick = ticks;
427 td->td_ru.ru_nivcsw++;
428 td->td_swinvoltick = ticks;
439 td->td_runtime += runtime;
440 td->td_incruntime += runtime;
442 td->td_generation++; /* bump preempt-detect counter */
446 td->td_tid, td_get_sched(td), td->td_proc->p_pid, td->td_name);
452 sched_switch(td, newtd, flags);
454 td->td_tid, td_get_sched(td), td->td_proc->p_pid, td->td_name);
459 if ((td = PCPU_GET(deadthread))) {
461 thread_stash(td);
471 setrunnable(struct thread *td)
474 THREAD_LOCK_ASSERT(td, MA_OWNED);
475 KASSERT(td->td_proc->p_state != PRS_ZOMBIE,
476 ("setrunnable: pid %d is a zombie", td->td_proc->p_pid));
477 switch (td->td_state) {
486 if (td->td_inhibitors != TDI_SWAPPED)
492 printf("state is 0x%x", td->td_state);
495 if ((td->td_flags & TDF_INMEM) == 0) {
496 if ((td->td_flags & TDF_SWAPINREQ) == 0) {
497 td->td_flags |= TDF_SWAPINREQ;
501 sched_wakeup(td);
560 struct thread *td;
562 td = curthread;
564 thread_lock(td);
566 prio = td->td_user_pri;
568 sched_prio(td, prio);
570 thread_unlock(td);
578 sys_yield(struct thread *td, struct yield_args *uap)
581 thread_lock(td);
582 if (PRI_BASE(td->td_pri_class) == PRI_TIMESHARE)
583 sched_prio(td, PRI_MAX_TIMESHARE);
585 thread_unlock(td);
586 td->td_retval[0] = 0;