/macosx-10.5.8/xnu-1228.15.4/osfmk/i386/commpage/ |
H A D | bzero_scalar.s | 77 pushl %ebx 95 movl %ecx,%ebx 96 subl %edx,%ebx 102 movl %ebx,%ecx /* zero by words */ 107 movl %ebx,%ecx 112 popl %ebx
|
H A D | longcopy_sse3x.s | 63 pushl %ebx // we'll need to use this too 67 movl %edi,%ebx // copy dest ptr 68 negl %ebx 69 andl $63,%ebx // get #bytes to cache line align destination 74 pushl %ebx // arg3 - #bytes to align destination (1..63) 83 addl %ebx,%esi // adjust ptrs and lengths past copy 84 addl %ebx,%edi 85 subl %ebx,%ecx 107 // ebx = scratch reg used to read a byte of each cache line 112 movzb (%esi,%eax),%ebx // touc [all...] |
H A D | commpage_mach_absolute_time.s | 52 pushl %ebx 68 movl %edx,%ebx 70 movl %ebx,%eax 71 movl %edx,%ebx 73 addl %ebx,%eax 82 popl %ebx 96 push %ebx 116 mov %edx, %ebx 120 mov %ebx, %eax 132 mov %eax, %ebx [all...] |
H A D | bcopy_sse2.s | 227 pushl %ebx // we'll need to use this 228 movl %edi,%ebx // copy dest ptr 229 negl %ebx 230 andl $63,%ebx // get #bytes to cache line align destination 236 pushl %ebx // arg3 - #bytes to align destination (1..63) 242 addl %ebx,%esi // adjust ptrs and lengths past copy 243 addl %ebx,%edi 244 subl %ebx,%ecx 264 // ebx = scratch reg used to read a byte of each cache line 269 movzb (%esi,%eax),%ebx // touc [all...] |
H A D | commpage_gettimeofday.s | 43 push %ebx 56 mov _COMM_PAGE_GTOD_SEC_BASE,%ebx /* load all the data before checking generation */ 63 add %eax,%ebx 71 mov %ebx,(%ecx) 76 pop %ebx
|
H A D | atomic.s | 74 // It assumes old -> %eax/%edx, new -> %ebx/%ecx, value -> %esi 188 pushl %ebx 190 movl 20(%esp),%ebx // %ebx == new 195 movl %eax,(%ebx,%esi)// link to old list head from new element 201 popl %ebx 214 pushl %ebx 222 movl (%eax,%esi),%ebx // point to 2nd in Q 229 popl %ebx
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/i386/ |
H A D | hibernate_restore.s | 61 * ebx = index into page table 67 shll $3,%ebx ; \ 68 addl base,%ebx ; \ 72 1: movl %eax,(%ebx) ; /* low 32b */ \ 73 addl $4,%ebx ; \ 74 movl %edx,(%ebx) ; /* high 32b */ \ 76 addl $4,%ebx ; /* next entry */ \ 134 xorl %ebx, %ebx 138 movl $(HPDPT), %ebx [all...] |
H A D | cswitch.s | 107 movl %ebx,KSS_EBX(%ecx) /* save registers */ 115 movl 8(%esp),%ebx /* get new thread */ 116 movl %ebx,%gs:CPU_ACTIVE_THREAD /* new thread is active */ 117 movl TH_KERNEL_STACK(%ebx),%ecx /* get its kernel stack */ 118 lea KERNEL_STACK_SIZE-IKS_SIZE-IEL_SIZE(%ecx),%ebx 122 movl %ebx,%gs:CPU_KERNEL_STACK /* set stack top */ 129 movl KSS_EBX(%ecx),%ebx 135 call *%ebx /* call real continuation */ 150 movl %ebx,KSS_EBX(%ecx) /* save registers */ 159 movl 4(%esp),%ebx /* ge [all...] |
H A D | postcode.h | 75 * Entry: value in %ebx 76 * Exit: %ebx preserved; %eax destroyed 79 roll $8, %ebx; \ 80 movl %ebx, %eax; \ 83 roll $8, %ebx; \ 84 movl %ebx, %eax; \ 87 roll $8, %ebx; \ 88 movl %ebx, %eax; \ 91 roll $8, %ebx; \ 92 movl %ebx, [all...] |
H A D | locore.s | 238 * Uses %eax, %ebx, %ecx, %edx, %esi, %edi. 262 movl %gs:CPU_PROCESSOR,%ebx /* get current processor */ ; \ 263 movl THREAD_TIMER(%ebx),%ecx /* get current timer */ ; \ 270 movl %ecx,THREAD_TIMER(%ebx) /* set current timer */ ; \ 273 movl CURRENT_STATE(%ebx),%ecx /* current state */ ; \ 278 movl %ecx,CURRENT_STATE(%ebx) /* set current state */ ; \ 284 * Uses %eax,%ebx,%ecx,%edx,%esi,%edi. 290 * Uses %eax,%ebx,%ecx,%edx,%esi,%edi. 296 * Uses %eax,%ebx,%ecx,%edx,%esi,%edi. 305 movl %gs:CPU_PROCESSOR,%ebx /* ge [all...] |
H A D | gdb_defs.h | 47 unsigned int ebx; member in struct:__anon704
|
H A D | start.s | 223 * ebx = index into page table 229 shll $(PTEINDX),%ebx ; \ 230 addl base,%ebx ; \ 233 1: movl %eax,(%ebx) ; \ 235 addl $(PTESIZE),%ebx ; /* next pte */ \ 245 movl %eax, %ebx ; \ 246 shrl $(PAGE_SHIFT), %ebx ; \ 267 movl KADDR(%ebp), %ebx // Load boot image phys addr 268 movl %ebx, %edx // Set edx with boot load phys addr 289 movl %ebx, [all...] |
H A D | rtclock.h | 92 movl %edx,%ebx ; \ 94 movl %ebx,%eax ; \ 95 movl %edx,%ebx ; \ 97 addl %ebx,%eax ; \
|
H A D | machine_routines_asm.s | 91 pushl %ebx // Save a volatile 115 movl %eax,%ebx // Need to save the low order part 121 // Now have %ecx:%ebx as low part of high low and %edx:%eax as high part of high high 124 addl %edi,%ebx // Add the low words 128 movl %ebx,%eax // Pass back low word 133 popl %ebx // Restore a volatile 216 pushl %ebx 225 popl %ebx 251 mov %edx, %ebx 255 mov %ebx, [all...] |
H A D | _setjmp.s | 75 movl %ebx,0(%ecx) 88 movl 0(%ecx),%ebx
|
H A D | asm.h | 198 * Gpush saves the %ebx register used for the GOT address 199 * Gpop pops %ebx if we need a GOT 200 * Gload loads %ebx with the GOT address if shared libraries are used 228 #define Gpush pushl %ebx 229 #define Gpop popl %ebx 230 #define Gload call 9f; 9: popl %ebx; addl $_GLOBAL_OFFSET_TABLE_+[.-9b],%ebx 232 #define Gotoff(lab) lab@GOTOFF(%ebx) 233 #define Gotoff2(l,r) l@GOTOFF(%ebx,r) 234 #define Gotoff3(l,r,s) l@GOTOFF(%ebx, [all...] |
H A D | idt.s | 106 movl $ EXT(lo_alltraps),%ebx ;\ 119 movl $ EXT(lo_alltraps),%ebx ;\ 159 movl $ EXT(lo_alltraps),%ebx ;\ 173 movl $ EXT(lo_allintrs),%ebx ;\ 487 * ebx 502 movl %esp,%ebx 504 subl ACT_PCB_ISS(%ecx),%ebx 520 addl %gs:CPU_HI_ISS,%ebx /* rebase PCB save area to high addr */ 523 movl %ebx,%esp /* switch to hi based PCB stack */ 564 movl $ EXT(lo_unix_scall),%ebx [all...] |
H A D | mp_slave_boot.s | 161 movl $0x28, %ebx /* segment */ 162 push %ebx 166 movl $0x20, %ebx
|
H A D | cpuid.c | 130 cache_linesize = bitfield(reg[ebx], 11, 0) + 1; 131 cache_partitions = bitfield(reg[ebx], 21, 12) + 1; 132 cache_associativity = bitfield(reg[ebx], 31, 22) + 1; 297 bcopy((char *)&cpuid_reg[ebx], &info_p->cpuid_vendor[0], 4); /* ug */ 358 info_p->cpuid_brand = bitfield(cpuid_reg[ebx], 7, 0); 369 bitfield(cpuid_reg[ebx], 23, 16); 406 info_p->cpuid_mwait_linesize_max = cpuid_reg[ebx]; 419 bitfield(cpuid_reg[ebx], 3, 0); 436 cpuid_reg[ebx]; 722 0, cpid[eax], cpid[ebx], cpi [all...] |
H A D | tss.h | 83 uint32_t ebx; member in struct:i386_tss
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/profiling/i386/ |
H A D | profile-asm.s | 664 * Loadup %ebx with the address of _profile_vars. On a multiprocessor, this 682 #define Vload CPU_NUMBER(%ebx); movl EXT(_profile_vars_cpus)(,%ebx,4),%ebx 685 #define Vload Gload; Egaddr(%ebx,_profile_vars) 694 * %ebx contains the address of the appropriate profile_vars structure. 711 xchgb %cl,V_recursive_alloc(%ebx) 721 leal V_acontext(%ebx,%edi,4),%ecx 724 /* %eax = size, %ebx = vars addr, %ecx = ptr to allocation context to try */ 738 SDADDNEG(%edx,V_acontext_locked(%ebx)) /* incremen [all...] |
/macosx-10.5.8/xnu-1228.15.4/libkern/i386/ |
H A D | OSAtomic.s | 53 pushl %ebx 57 movl 12+8(%esp), %ebx #; low 32-bits of newValue 61 cmpxchg8b 0(%edi) #; CAS (eax:edx, ebx:ecx implicit) 65 popl %ebx 76 pushl %ebx 82 movl %eax, %ebx 83 movl %edx, %ecx #; ebx:ecx := *ptr 84 addl 4+8(%esp), %ebx 85 adcl 8+8(%esp), %ecx #; ebx:ecx := *ptr + theAmount 87 cmpxchg8b 0(%edi) #; CAS (eax:edx, ebx [all...] |
/macosx-10.5.8/xnu-1228.15.4/bsd/crypto/aes/i386/ |
H A D | aes_x86_v2.s | 74 * interface as my C implementation. The ebx, esi, edi and ebp registers are 226 roll $16, %ebx; \ 236 shrl $16, %ebx; \ 238 orl %ebx, %eax; \ 241 m1(ebp, ah, 1, ebx); \ 242 m1(ebp, dh, 3, ebx); \ 243 m2(ebx, dl, 2, ebx); \ 244 m1(ebx, ch, 1, edx); \ 245 ## m1 ## _zo(ebx, a [all...] |
/macosx-10.5.8/xnu-1228.15.4/libsa/i386/ |
H A D | setjmp.s | 78 movl %ebx,0(%ecx) 92 movl 0(%ecx),%ebx
|
/macosx-10.5.8/xnu-1228.15.4/EXTERNAL_HEADERS/architecture/i386/ |
H A D | tss.h | 63 unsigned int ebx; member in struct:tss
|