• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.5.8/xnu-1228.15.4/osfmk/i386/

Lines Matching refs:edx

117 #define	M_ILK		(%edx)
118 #define M_LOCKED MUTEX_LOCKED(%edx)
119 #define M_WAITERS MUTEX_WAITERS(%edx)
120 #define M_PROMOTED_PRI MUTEX_PROMOTED_PRI(%edx)
121 #define M_ITAG MUTEX_ITAG(%edx)
122 #define M_PTR MUTEX_PTR(%edx)
124 #define M_TYPE MUTEX_TYPE(%edx)
125 #define M_PC MUTEX_PC(%edx)
126 #define M_THREAD MUTEX_THREAD(%edx)
328 movl L_ARG0,%edx /* fetch lock pointer */
329 movl $0,(%edx) /* clear the lock */
339 movl L_ARG0,%edx /* fetch lock pointer */
340 movb $0,(%edx) /* clear the lock */
350 movl L_ARG0,%edx /* fetch lock pointer */
355 movl (%edx), %eax
358 lock; cmpxchgl %ecx,(%edx) /* try to acquire the HW lock */
374 movl L_ARG0,%edx /* Load lock pointer */
378 movb (%edx), %al /* Load byte at address */
381 lock; cmpxchgb %cl,(%edx) /* attempt atomic compare exchange */
396 movl L_ARG0,%edx /* fetch lock pointer */
403 movl (%edx), %eax
406 lock; cmpxchgl %ecx,(%edx) /* try to acquire the HW lock */
421 mov %edx,%edi
424 rdtsc /* read cyclecount into %edx:%eax */
427 adcl $0,%edx /* add carry */
428 mov %edx,%ecx
434 mov $(INNER_LOOP_COUNT),%edx
440 decl %edx /* decrement inner loop count */
447 rdtsc /* cyclecount into %edx:%eax */
449 cmpl %ecx,%edx /* compare high-order 32-bits */
463 movl %gs:CPU_ACTIVE_THREAD,%edx
464 lock; cmpxchgl %edx,(%edi) /* try to acquire the HW lock */
478 movl L_ARG0,%edx /* fetch lock pointer */
479 movl $0,(%edx) /* clear the lock */
490 movl L_ARG0,%edx /* Load lock pointer */
491 movb $0,(%edx) /* Clear the lock byte */
508 movl L_ARG0,%edx /* Fetch lock pointer */
509 movl $0,(%edx) /* Clear the lock */
511 clflush (%edx) /* Write back and invalidate line */
520 movl L_ARG0,%edx /* fetch lock pointer */
524 movl (%edx),%eax
527 lock; cmpxchgl %ecx,(%edx) /* try to acquire the HW lock */
544 movl L_ARG0,%edx /* fetch lock pointer */
546 movl (%edx),%eax /* check lock value */
553 movl L_ARG0,%edx /* fetch lock pointer */
592 movl S_ARG0, %edx
594 movl (%edx), %eax /* Load state bitfield and interlock */
600 cmpxchgl %ecx, (%edx) /* Attempt atomic exchange */
611 /* Fall thru when patched, counting on lock pointer in %edx */
612 LOCKSTAT_RECORD(LS_LCK_RW_LOCK_SHARED_ACQUIRE, %edx)
636 movl S_ARG0, %edx
638 movl (%edx), %eax /* Load state bitfield and interlock */
650 cmpxchgl %ecx, (%edx)
660 * %edx contains the lock address already from the above
662 LOCKSTAT_RECORD2(LS_LCK_RW_DONE_RELEASE, %edx, $0)
676 movl B_ARG0,%edx /* fetch lock pointer */
711 /* %edx contains the lock address from above */
712 LOCKSTAT_RECORD(LS_MUTEX_LOCK_SPIN_ACQUIRE, %edx)
730 movl B_ARG0,%edx /* fetch lock pointer */
771 /* %edx still contains the lock pointer */
772 LOCKSTAT_RECORD(LS_MUTEX_LOCK_ACQUIRE, %edx)
781 pushl %edx /* save mutex address */
782 pushl %edx
785 popl %edx /* restore mutex address */
821 push %edx /* lock address */
824 movl B_ARG0,%edx /* refetch mutex address */
845 pushl %edx /* push mutex address */
848 movl B_ARG0,%edx /* refetch mutex address */
866 movl B_ARG0,%edx /* fetch lock pointer */
902 /* %edx inherits the lock pointer from above */
903 LOCKSTAT_RECORD(LS_MUTEX_TRY_SPIN_ACQUIRE, %edx)
932 movl B_ARG0,%edx /* fetch lock pointer */
973 /* inherit the lock pointer in %edx from above */
974 LOCKSTAT_RECORD(LS_MUTEX_TRY_LOCK_ACQUIRE, %edx)
980 pushl %edx /* save mutex address */
981 pushl %edx
984 popl %edx /* restore mutex address */
1019 movl L_ARG0,%edx /* fetch lock pointer */
1039 /* inherit %edx from above */
1040 LOCKSTAT_RECORD(LS_MUTEX_CONVERT_SPIN_ACQUIRE, %edx)
1046 pushl %edx /* save mutex address */
1047 pushl %edx
1050 popl %edx /* restore mutex address */
1056 movl B_ARG0,%edx /* fetch lock pointer */
1074 /* inherit lock pointer in %edx from above */
1075 LOCKSTAT_RECORD(LS_MUTEX_UNLOCK_RELEASE, %edx)
1080 pushl %edx /* save mutex address */
1081 pushl %edx /* push mutex address */
1084 popl %edx /* restore mutex pointer */
1119 /* inherit %edx from above */
1120 LOCKSTAT_RECORD(LS_MUTEX_UNLOCK_RELEASE, %edx)
1137 pushl %edx /* push mutex address */
1140 movl B_ARG0,%edx /* restore lock pointer */
1156 movl S_ARG0,%edx /* Load lock address */
1160 cmove M_PTR,%edx /* If so, take indirection */
1175 movl %edx,S_ARG1 /* Prep assertion failure */
1179 movl %edx,S_ARG1 /* Prep assertion failure */
1218 movl B_ARG0,%edx /* fetch lock pointer */
1258 pushl %edx
1263 popl %edx
1269 cmove M_PTR,%edx /* If so, take indirection */
1275 movl MUTEX_GRP(%edx),%esi /* Load lock group */
1319 pushl %edx
1324 popl %edx
1331 movl B_ARG0,%edx /* fetch lock pointer */
1366 /* inherit lock pointer in %edx above */
1367 LOCKSTAT_RECORD(LS_LCK_MTX_LOCK_ACQUIRE, %edx)
1372 pushl %edx /* save mutex address */
1373 pushl %edx
1376 popl %edx /* restore mutex address */
1393 pushl %edx
1398 popl %edx
1421 pushl %edx /* save mutex address */
1422 pushl %edx
1425 popl %edx /* restore mutex address */
1445 pushl %edx /* save mutex address */
1447 pushl %edx /* push mutex address */
1454 popl %edx /* restore mutex address */
1470 pushl %edx
1475 popl %edx
1481 cmove M_PTR,%edx /* If so, take indirection */
1507 movl MUTEX_GRP(%edx),%esi /* Load lock group */
1547 /* inherit lock pointer in %edx above */
1548 LOCKSTAT_RECORD(LS_LCK_MTX_EXT_LOCK_ACQUIRE, %edx)
1553 pushl %edx /* save mutex address */
1554 pushl %edx
1557 popl %edx /* restore mutex address */
1575 pushl %edx
1580 popl %edx
1625 pushl %edx /* save mutex address */
1626 pushl %edx
1629 popl %edx /* restore mutex address */
1656 pushl %edx /* save mutex address */
1658 pushl %edx /* push mutex address */
1665 popl %edx /* restore mutex address */
1682 pushl %edx
1687 popl %edx
1694 movl B_ARG0,%edx /* fetch lock pointer */
1724 /* inherit lock pointer in %edx above */
1725 LOCKSTAT_RECORD(LS_LCK_MTX_TRY_SPIN_LOCK_ACQUIRE, %edx)
1752 pushl %edx
1757 popl %edx
1762 cmove M_PTR,%edx /* If so, take indirection */
1769 movl MUTEX_GRP(%edx),%esi /* Load lock group */
1783 movl B_ARG0,%edx /* fetch lock pointer */
1821 /* inherit lock pointer in %edx from above */
1822 LOCKSTAT_RECORD(LS_LCK_MTX_TRY_LOCK_ACQUIRE, %edx)
1828 pushl %edx /* save mutex address */
1829 pushl %edx
1832 popl %edx /* restore mutex address */
1857 pushl %edx
1862 popl %edx
1872 cmpl %edx,B_ARG0
1881 movl MUTEX_GRP(%edx),%esi /* Load lock group */
1893 cmove M_PTR,%edx /* If so, take indirection */
1900 movl MUTEX_GRP(%edx),%esi /* Load lock group */
1914 movl L_ARG0,%edx /* fetch lock pointer */
1917 cmove M_PTR,%edx /* If so, take indirection */
1937 pushl %edx /* save mutex address */
1938 pushl %edx
1941 popl %edx /* restore mutex address */
1948 movl B_ARG0,%edx /* fetch lock pointer */
1951 cmove M_PTR,%edx /* If so, take indirection */
1971 /* inherit lock pointer in %edx from above */
1972 LOCKSTAT_RECORD(LS_LCK_MTX_UNLOCK_RELEASE, %edx)
1977 pushl %edx /* save mutex address */
1978 pushl %edx /* push mutex address */
1981 popl %edx /* restore mutex pointer */
2012 /* inherit lock pointer in %edx above */
2013 LOCKSTAT_RECORD(LS_LCK_MTX_UNLOCK_RELEASE, %edx)
2030 pushl %edx
2035 popl %edx
2039 pushl %edx /* save mutex address */
2041 pushl %edx /* push mutex address */
2044 popl %edx /* restore mutex pointer */
2057 /* inherit lock pointer in %edx from above */
2058 LOCKSTAT_RECORD(LS_LCK_MTX_EXT_UNLOCK_RELEASE, %edx)
2064 movl L_ARG0,%edx /* no indirection here */
2157 movl L_ARG0,%edx
2160 bts %edx,(%eax)
2164 movl L_ARG0,%edx
2167 btr %edx,(%eax)
2205 movl %eax, %edx
2208 addl %edx, %eax /* Calculate result */
2215 movl %eax, %edx
2218 addl %edx, %eax /* Calculate result */
2225 movl L_ARG1, %edx /* Load mask */
2226 orl %eax, %edx
2228 cmpxchgl %edx, (%ecx) /* Atomic CAS */
2230 movl %edx, %eax /* Result */
2239 movl L_ARG1, %edx /* Load mask */
2241 orl %edx, (%ecx) /* Atomic OR */
2248 movl L_ARG1, %edx /* Load mask */
2249 andl %eax, %edx
2251 cmpxchgl %edx, (%ecx) /* Atomic CAS */
2253 movl %edx, %eax /* Result */
2262 movl L_ARG1, %edx /* Load mask */
2264 andl %edx, (%ecx) /* Atomic OR */